Dec 03 00:23:33 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 00:23:33 crc restorecon[4751]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:33 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 00:23:34 crc restorecon[4751]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 00:23:34 crc kubenswrapper[4912]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 00:23:34 crc kubenswrapper[4912]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 00:23:34 crc kubenswrapper[4912]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 00:23:34 crc kubenswrapper[4912]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 00:23:34 crc kubenswrapper[4912]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 00:23:34 crc kubenswrapper[4912]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.449302 4912 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453745 4912 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453776 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453783 4912 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453789 4912 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453796 4912 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453801 4912 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453806 4912 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453811 4912 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453817 4912 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453822 4912 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453827 4912 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453832 4912 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453837 4912 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453842 4912 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453847 4912 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453853 4912 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453859 4912 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453868 4912 feature_gate.go:330] unrecognized feature gate: Example Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453873 4912 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453879 4912 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453884 4912 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453889 4912 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453894 4912 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453899 4912 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453905 4912 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453912 4912 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453920 4912 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453928 4912 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453934 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453941 4912 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453947 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453953 4912 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453960 4912 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453965 4912 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453971 4912 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453976 4912 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453981 4912 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453986 4912 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453991 4912 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.453999 4912 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454004 4912 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454010 4912 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454015 4912 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454020 4912 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454025 4912 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454030 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454035 4912 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454040 4912 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454045 4912 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454050 4912 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454055 4912 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454061 4912 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454066 4912 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454071 4912 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454076 4912 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454082 4912 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454087 4912 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454092 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454099 4912 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454104 4912 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454109 4912 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454114 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454119 4912 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454125 4912 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454130 4912 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454135 4912 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454140 4912 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454145 4912 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454150 4912 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454158 4912 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.454165 4912 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454318 4912 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454335 4912 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454361 4912 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454370 4912 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454381 4912 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454387 4912 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454396 4912 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454403 4912 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454410 4912 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454417 4912 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454424 4912 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454431 4912 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454457 4912 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454465 4912 flags.go:64] FLAG: --cgroup-root="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454471 4912 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454478 4912 flags.go:64] FLAG: --client-ca-file="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454485 4912 flags.go:64] FLAG: --cloud-config="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454491 4912 flags.go:64] FLAG: --cloud-provider="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454497 4912 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454505 4912 flags.go:64] FLAG: --cluster-domain="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454511 4912 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454517 4912 flags.go:64] FLAG: --config-dir="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454523 4912 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454530 4912 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454538 4912 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454544 4912 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454550 4912 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454557 4912 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454563 4912 flags.go:64] FLAG: --contention-profiling="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454569 4912 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454575 4912 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454581 4912 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454587 4912 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454595 4912 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454601 4912 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454606 4912 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454612 4912 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454618 4912 flags.go:64] FLAG: --enable-server="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454624 4912 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454632 4912 flags.go:64] FLAG: --event-burst="100" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454638 4912 flags.go:64] FLAG: --event-qps="50" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454644 4912 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454653 4912 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454661 4912 flags.go:64] FLAG: --eviction-hard="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454681 4912 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454687 4912 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454693 4912 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454700 4912 flags.go:64] FLAG: --eviction-soft="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454707 4912 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454713 4912 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454719 4912 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454725 4912 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454731 4912 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454737 4912 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454743 4912 flags.go:64] FLAG: --feature-gates="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454751 4912 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454757 4912 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454764 4912 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454770 4912 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454776 4912 flags.go:64] FLAG: --healthz-port="10248" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454782 4912 flags.go:64] FLAG: --help="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454788 4912 flags.go:64] FLAG: --hostname-override="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454794 4912 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454800 4912 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454807 4912 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454813 4912 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454818 4912 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454824 4912 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454830 4912 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454836 4912 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454842 4912 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454848 4912 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454855 4912 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454861 4912 flags.go:64] FLAG: --kube-reserved="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454867 4912 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454873 4912 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454879 4912 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454886 4912 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454892 4912 flags.go:64] FLAG: --lock-file="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454898 4912 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454904 4912 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454910 4912 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454919 4912 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454925 4912 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454932 4912 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454938 4912 flags.go:64] FLAG: --logging-format="text" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454943 4912 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454950 4912 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454956 4912 flags.go:64] FLAG: --manifest-url="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454962 4912 flags.go:64] FLAG: --manifest-url-header="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454971 4912 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454977 4912 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454984 4912 flags.go:64] FLAG: --max-pods="110" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454991 4912 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.454997 4912 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455003 4912 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455009 4912 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455015 4912 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455021 4912 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455027 4912 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455041 4912 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455048 4912 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455054 4912 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455060 4912 flags.go:64] FLAG: --pod-cidr="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455066 4912 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455075 4912 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455081 4912 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455087 4912 flags.go:64] FLAG: --pods-per-core="0" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455093 4912 flags.go:64] FLAG: --port="10250" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455099 4912 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455105 4912 flags.go:64] FLAG: --provider-id="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455111 4912 flags.go:64] FLAG: --qos-reserved="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455117 4912 flags.go:64] FLAG: --read-only-port="10255" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455124 4912 flags.go:64] FLAG: --register-node="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455130 4912 flags.go:64] FLAG: --register-schedulable="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455136 4912 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455145 4912 flags.go:64] FLAG: --registry-burst="10" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455151 4912 flags.go:64] FLAG: --registry-qps="5" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455157 4912 flags.go:64] FLAG: --reserved-cpus="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455164 4912 flags.go:64] FLAG: --reserved-memory="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455171 4912 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455178 4912 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455184 4912 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455190 4912 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455196 4912 flags.go:64] FLAG: --runonce="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455202 4912 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455208 4912 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455214 4912 flags.go:64] FLAG: --seccomp-default="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455220 4912 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455226 4912 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455232 4912 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455239 4912 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455245 4912 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455251 4912 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455257 4912 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455263 4912 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455270 4912 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455276 4912 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455283 4912 flags.go:64] FLAG: --system-cgroups="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455289 4912 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455299 4912 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455305 4912 flags.go:64] FLAG: --tls-cert-file="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455311 4912 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455319 4912 flags.go:64] FLAG: --tls-min-version="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455325 4912 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455331 4912 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455342 4912 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455348 4912 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455354 4912 flags.go:64] FLAG: --v="2" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455362 4912 flags.go:64] FLAG: --version="false" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455370 4912 flags.go:64] FLAG: --vmodule="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455377 4912 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455385 4912 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455563 4912 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455571 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455577 4912 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455583 4912 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455588 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455594 4912 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455599 4912 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455604 4912 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455609 4912 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455617 4912 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455624 4912 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455630 4912 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455636 4912 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455642 4912 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455648 4912 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455655 4912 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455661 4912 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455667 4912 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455673 4912 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455678 4912 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455684 4912 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455691 4912 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455698 4912 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455705 4912 feature_gate.go:330] unrecognized feature gate: Example Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455713 4912 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455721 4912 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455727 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455733 4912 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455738 4912 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455743 4912 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455748 4912 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455753 4912 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455759 4912 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455764 4912 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455769 4912 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455774 4912 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455779 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455785 4912 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455792 4912 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455797 4912 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455802 4912 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455808 4912 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455813 4912 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455818 4912 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455823 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455829 4912 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455834 4912 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455840 4912 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455845 4912 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455850 4912 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455855 4912 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455861 4912 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455866 4912 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455871 4912 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455877 4912 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455882 4912 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455890 4912 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455897 4912 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455903 4912 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455908 4912 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455913 4912 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455918 4912 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455924 4912 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455929 4912 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455934 4912 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455940 4912 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455945 4912 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455950 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455955 4912 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455961 4912 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.455966 4912 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.455975 4912 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.462455 4912 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.462495 4912 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462568 4912 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462574 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462580 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462584 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462588 4912 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462593 4912 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462597 4912 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462601 4912 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462605 4912 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462610 4912 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462616 4912 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462620 4912 feature_gate.go:330] unrecognized feature gate: Example Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462625 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462629 4912 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462634 4912 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462638 4912 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462642 4912 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462646 4912 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462650 4912 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462654 4912 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462657 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462661 4912 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462664 4912 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462668 4912 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462671 4912 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462675 4912 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462679 4912 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462685 4912 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462689 4912 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462693 4912 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462697 4912 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462701 4912 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462706 4912 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462709 4912 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462715 4912 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462718 4912 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462722 4912 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462727 4912 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462732 4912 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462736 4912 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462740 4912 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462744 4912 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462747 4912 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462751 4912 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462755 4912 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462759 4912 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462762 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462766 4912 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462770 4912 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462774 4912 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462777 4912 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462781 4912 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462784 4912 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462788 4912 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462792 4912 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462795 4912 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462799 4912 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462802 4912 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462807 4912 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462810 4912 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462814 4912 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462818 4912 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462821 4912 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462825 4912 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462828 4912 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462832 4912 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462836 4912 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462840 4912 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462843 4912 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462847 4912 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462851 4912 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.462858 4912 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462970 4912 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462975 4912 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462980 4912 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462984 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462988 4912 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462991 4912 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462995 4912 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.462998 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463001 4912 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463005 4912 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463009 4912 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463012 4912 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463016 4912 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463020 4912 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463023 4912 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463027 4912 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463030 4912 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463034 4912 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463037 4912 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463041 4912 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463045 4912 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463048 4912 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463051 4912 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463055 4912 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463058 4912 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463062 4912 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463065 4912 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463069 4912 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463072 4912 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463076 4912 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463080 4912 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463083 4912 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463088 4912 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463092 4912 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463097 4912 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463101 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463105 4912 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463108 4912 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463111 4912 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463115 4912 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463118 4912 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463122 4912 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463126 4912 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463130 4912 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463134 4912 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463138 4912 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463142 4912 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463145 4912 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463149 4912 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463152 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463156 4912 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463160 4912 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463164 4912 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463170 4912 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463174 4912 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463178 4912 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463183 4912 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463188 4912 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463192 4912 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463195 4912 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463199 4912 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463203 4912 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463207 4912 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463211 4912 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463215 4912 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463219 4912 feature_gate.go:330] unrecognized feature gate: Example Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463222 4912 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463226 4912 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463229 4912 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463233 4912 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.463237 4912 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.463244 4912 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.463381 4912 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.465848 4912 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.465918 4912 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.466352 4912 server.go:997] "Starting client certificate rotation" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.466377 4912 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.466684 4912 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-26 02:15:15.076785478 +0000 UTC Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.466776 4912 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 553h51m40.610012352s for next certificate rotation Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.470367 4912 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.471768 4912 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.479162 4912 log.go:25] "Validated CRI v1 runtime API" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.497702 4912 log.go:25] "Validated CRI v1 image API" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.499415 4912 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.501380 4912 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-00-19-19-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.501405 4912 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.513825 4912 manager.go:217] Machine: {Timestamp:2025-12-03 00:23:34.51280044 +0000 UTC m=+0.154821010 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:92184f41-123a-4ffc-8bda-9fa272ab6757 BootID:bc17af26-6c16-455e-9b60-e30600d665e4 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:9a:eb:bf Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:9a:eb:bf Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:10:cf:96 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:58:82:dd Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:91:d3:ca Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:bb:36:1d Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:c0:00:4f Speed:-1 Mtu:1496} {Name:eth10 MacAddress:4a:8b:80:04:2c:77 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:62:fc:be:2b:df:2f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.514007 4912 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.514142 4912 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.514377 4912 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.514576 4912 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.514605 4912 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.514782 4912 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.514792 4912 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.514991 4912 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.515024 4912 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.515255 4912 state_mem.go:36] "Initialized new in-memory state store" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.515599 4912 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.516620 4912 kubelet.go:418] "Attempting to sync node with API server" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.516643 4912 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.516666 4912 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.516676 4912 kubelet.go:324] "Adding apiserver pod source" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.516686 4912 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.517976 4912 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.518280 4912 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519239 4912 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.519205 4912 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.519235 4912 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.519285 4912 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.519317 4912 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519724 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519752 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519760 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519767 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519778 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519784 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519790 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519800 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519807 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519814 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519830 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.519837 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.520013 4912 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.520411 4912 server.go:1280] "Started kubelet" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.520659 4912 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.521192 4912 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.521213 4912 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.521748 4912 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 00:23:34 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.525110 4912 server.go:460] "Adding debug handlers to kubelet server" Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.526657 4912 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.20:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d8cc6d0034989 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 00:23:34.520392073 +0000 UTC m=+0.162412633,LastTimestamp:2025-12-03 00:23:34.520392073 +0000 UTC m=+0.162412633,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.528693 4912 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.528744 4912 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.528953 4912 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.528958 4912 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 01:01:50.336321616 +0000 UTC Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.528962 4912 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.529010 4912 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.528994 4912 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 480h38m15.807329701s for next certificate rotation Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.528982 4912 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.529562 4912 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.529658 4912 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.530187 4912 factory.go:55] Registering systemd factory Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.530219 4912 factory.go:221] Registration of the systemd container factory successfully Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.530578 4912 factory.go:153] Registering CRI-O factory Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.530607 4912 factory.go:221] Registration of the crio container factory successfully Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.530677 4912 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.530698 4912 factory.go:103] Registering Raw factory Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.530712 4912 manager.go:1196] Started watching for new ooms in manager Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.530581 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="200ms" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.531248 4912 manager.go:319] Starting recovery of all containers Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540006 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540075 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540095 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540115 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540157 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540174 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540192 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540209 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540229 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540246 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540263 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540279 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540308 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540350 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540385 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540413 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540477 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540500 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540518 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540538 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540555 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540571 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540590 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540607 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540624 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540642 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540666 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540686 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540713 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540732 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540748 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540765 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540785 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.540803 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542124 4912 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542200 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542225 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542248 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542262 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542281 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542301 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542317 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542333 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542376 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542391 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542407 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542421 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542459 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542480 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542493 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542511 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542524 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542538 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542559 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542579 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542595 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542611 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542629 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542641 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542654 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542667 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542686 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542703 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542717 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542737 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542750 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542770 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542789 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542808 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542824 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542854 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542871 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542884 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542901 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542915 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542928 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542946 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542961 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542981 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.542995 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543015 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543029 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543041 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543053 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543065 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543102 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543115 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543126 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543148 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543160 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543425 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543465 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543479 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543495 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543512 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543523 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543539 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543552 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543580 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543595 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543607 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543618 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543629 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543641 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543652 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543669 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543682 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543699 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543716 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543733 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.543748 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546267 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546306 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546331 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546356 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546377 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546397 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546416 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546465 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546489 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546509 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546530 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546548 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546569 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546592 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546611 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546633 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546652 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546671 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546700 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546718 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546738 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546751 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546766 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546779 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546793 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546807 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546820 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546839 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546853 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546866 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546879 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546892 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546905 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546920 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546940 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.546989 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547021 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547052 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547087 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547112 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547131 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547157 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547184 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547208 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547226 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547342 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547375 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547400 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547417 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547457 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547474 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547499 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547516 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547533 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547551 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547569 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547593 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547611 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547632 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547656 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547675 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547693 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547713 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547741 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547760 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547781 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547801 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547834 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547852 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.547871 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550105 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550237 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550284 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550331 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550362 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550495 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550511 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550541 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550584 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550601 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550617 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550633 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550666 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550681 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550696 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550706 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550735 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550755 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550766 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550778 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550812 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550825 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550841 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550854 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550865 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550894 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550904 4912 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550914 4912 reconstruct.go:97] "Volume reconstruction finished" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.550923 4912 reconciler.go:26] "Reconciler: start to sync state" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.559823 4912 manager.go:324] Recovery completed Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.567868 4912 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.570211 4912 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.570261 4912 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.570295 4912 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.570368 4912 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.570843 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: W1203 00:23:34.572092 4912 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.572168 4912 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.572195 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.572220 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.572231 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.573407 4912 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.573420 4912 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.573449 4912 state_mem.go:36] "Initialized new in-memory state store" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.586081 4912 policy_none.go:49] "None policy: Start" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.587575 4912 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.587593 4912 state_mem.go:35] "Initializing new in-memory state store" Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.629334 4912 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.632586 4912 manager.go:334] "Starting Device Plugin manager" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.632628 4912 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.632638 4912 server.go:79] "Starting device plugin registration server" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.633049 4912 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.633067 4912 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.633193 4912 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.633274 4912 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.633288 4912 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.638564 4912 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.671031 4912 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.671180 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.672154 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.672196 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.672205 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.672393 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.672517 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.672635 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.673209 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.673261 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.673274 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.673385 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.673474 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.673507 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.673519 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.673882 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.673939 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.674121 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.674146 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.674198 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.674335 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.674422 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.674474 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.674597 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.674637 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.674649 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.675318 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.675340 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.675367 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.675389 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.675408 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.675419 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.675533 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.675622 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.675641 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.676422 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.676478 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.676494 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.676666 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.676697 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.676705 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.676782 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.676819 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.677631 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.677667 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.677680 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.731628 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="400ms" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.733191 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.735529 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.735572 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.735587 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.735612 4912 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.736500 4912 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.753825 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.753857 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.753877 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.753895 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.753911 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.753950 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.753990 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.754028 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.754061 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.754078 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.754127 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.754147 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.754241 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.754268 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.754289 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.855662 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.855987 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856049 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856016 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856261 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856330 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856396 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856529 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856614 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856579 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856583 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856540 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856635 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856662 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.856971 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857127 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857209 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857244 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857303 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857333 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857387 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857478 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857517 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857540 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857618 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857648 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857743 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857716 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857901 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.857658 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.937171 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.938868 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.938947 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.938965 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.939003 4912 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:23:34 crc kubenswrapper[4912]: E1203 00:23:34.939754 4912 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Dec 03 00:23:34 crc kubenswrapper[4912]: I1203 00:23:34.995712 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.001630 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.017878 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:35 crc kubenswrapper[4912]: W1203 00:23:35.029255 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-9cbc75208989b27965b96a7a88a4fb747a6bf0c1721fcc29d9db43ec8505eac3 WatchSource:0}: Error finding container 9cbc75208989b27965b96a7a88a4fb747a6bf0c1721fcc29d9db43ec8505eac3: Status 404 returned error can't find the container with id 9cbc75208989b27965b96a7a88a4fb747a6bf0c1721fcc29d9db43ec8505eac3 Dec 03 00:23:35 crc kubenswrapper[4912]: W1203 00:23:35.030819 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-569701a7fab72b3680929c0dd20872c64b26143691b1add75c0d2bc0e84e0b3f WatchSource:0}: Error finding container 569701a7fab72b3680929c0dd20872c64b26143691b1add75c0d2bc0e84e0b3f: Status 404 returned error can't find the container with id 569701a7fab72b3680929c0dd20872c64b26143691b1add75c0d2bc0e84e0b3f Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.033895 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:35 crc kubenswrapper[4912]: W1203 00:23:35.036963 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-cd2802feb9c7e85bdd533a1394f29eaf85ee46461edbaf8d7419efbe2195b7ec WatchSource:0}: Error finding container cd2802feb9c7e85bdd533a1394f29eaf85ee46461edbaf8d7419efbe2195b7ec: Status 404 returned error can't find the container with id cd2802feb9c7e85bdd533a1394f29eaf85ee46461edbaf8d7419efbe2195b7ec Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.041110 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:23:35 crc kubenswrapper[4912]: W1203 00:23:35.112113 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-1e2394db5ff9f2ad0bd2016745b80108a69fa10d648d0478df06cb90ee413dcb WatchSource:0}: Error finding container 1e2394db5ff9f2ad0bd2016745b80108a69fa10d648d0478df06cb90ee413dcb: Status 404 returned error can't find the container with id 1e2394db5ff9f2ad0bd2016745b80108a69fa10d648d0478df06cb90ee413dcb Dec 03 00:23:35 crc kubenswrapper[4912]: E1203 00:23:35.133328 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="800ms" Dec 03 00:23:35 crc kubenswrapper[4912]: W1203 00:23:35.134536 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-2defabddb723b012b50e0ae32ee7be248b0d5fedb3c9b8c4a45f8da7d5842bd8 WatchSource:0}: Error finding container 2defabddb723b012b50e0ae32ee7be248b0d5fedb3c9b8c4a45f8da7d5842bd8: Status 404 returned error can't find the container with id 2defabddb723b012b50e0ae32ee7be248b0d5fedb3c9b8c4a45f8da7d5842bd8 Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.340883 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.343383 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.343419 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.343429 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.343466 4912 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:23:35 crc kubenswrapper[4912]: E1203 00:23:35.343874 4912 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Dec 03 00:23:35 crc kubenswrapper[4912]: W1203 00:23:35.436310 4912 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:35 crc kubenswrapper[4912]: E1203 00:23:35.436774 4912 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.521741 4912 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.577238 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.577486 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1e2394db5ff9f2ad0bd2016745b80108a69fa10d648d0478df06cb90ee413dcb"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.579216 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88" exitCode=0 Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.579274 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.579369 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cd2802feb9c7e85bdd533a1394f29eaf85ee46461edbaf8d7419efbe2195b7ec"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.579529 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.580802 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.580860 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.580875 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.581120 4912 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832" exitCode=0 Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.581193 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.581223 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"569701a7fab72b3680929c0dd20872c64b26143691b1add75c0d2bc0e84e0b3f"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.581402 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.582523 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.582565 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.582577 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.582851 4912 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86" exitCode=0 Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.582918 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.582966 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9cbc75208989b27965b96a7a88a4fb747a6bf0c1721fcc29d9db43ec8505eac3"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.583049 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.583984 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.584099 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.584115 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.584875 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.587645 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.587674 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.587686 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.588859 4912 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3" exitCode=0 Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.588899 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.588944 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2defabddb723b012b50e0ae32ee7be248b0d5fedb3c9b8c4a45f8da7d5842bd8"} Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.589062 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.589784 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.589811 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:35 crc kubenswrapper[4912]: I1203 00:23:35.589820 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:35 crc kubenswrapper[4912]: W1203 00:23:35.643963 4912 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:35 crc kubenswrapper[4912]: E1203 00:23:35.644056 4912 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:23:35 crc kubenswrapper[4912]: W1203 00:23:35.842135 4912 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:35 crc kubenswrapper[4912]: E1203 00:23:35.842218 4912 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:23:35 crc kubenswrapper[4912]: E1203 00:23:35.934829 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="1.6s" Dec 03 00:23:35 crc kubenswrapper[4912]: W1203 00:23:35.987523 4912 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:35 crc kubenswrapper[4912]: E1203 00:23:35.987611 4912 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.144517 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.146490 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.146543 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.146552 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.146579 4912 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:23:36 crc kubenswrapper[4912]: E1203 00:23:36.147041 4912 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.521558 4912 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.593173 4912 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01" exitCode=0 Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.593266 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.593453 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.594269 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.594292 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.594300 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.596767 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"089195fc4e809a322be7b9f5a536d7fb8924fad3a5ad89ec6dd68994dd0e9646"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.596842 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.597485 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.597575 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.597596 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.599287 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.599315 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.599327 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.599403 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.600122 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.600145 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.600156 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.602169 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.602223 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.602209 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.602335 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.602874 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.602904 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.602914 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.605102 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.605126 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.605138 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.605150 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.605161 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0"} Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.605178 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.605714 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.605737 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:36 crc kubenswrapper[4912]: I1203 00:23:36.605746 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.609895 4912 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a" exitCode=0 Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.609970 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a"} Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.610076 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.610100 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.610209 4912 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.610257 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.611285 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.611294 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.611328 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.611344 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.611352 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.611373 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.611381 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.611302 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.611417 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.628659 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.747374 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.748396 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.748425 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.748467 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:37 crc kubenswrapper[4912]: I1203 00:23:37.748491 4912 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.616097 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e"} Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.616145 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d"} Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.616159 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4"} Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.616168 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63"} Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.616166 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.616229 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.616341 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038"} Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.617155 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.617180 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.617189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.617210 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.617235 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.617246 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.998736 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:38 crc kubenswrapper[4912]: I1203 00:23:38.998979 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.000270 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.000305 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.000317 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.001527 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.481887 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.598156 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.617866 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.617899 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.618921 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.618938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.618953 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.618957 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.618963 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:39 crc kubenswrapper[4912]: I1203 00:23:39.618967 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.229717 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.229854 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.230996 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.231030 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.231042 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.234397 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.394521 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.620069 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.620563 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.620638 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.621504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.621550 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.621604 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.623540 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.623575 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.623840 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.624016 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.624073 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:40 crc kubenswrapper[4912]: I1203 00:23:40.624092 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:41 crc kubenswrapper[4912]: I1203 00:23:41.622397 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:41 crc kubenswrapper[4912]: I1203 00:23:41.623135 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:41 crc kubenswrapper[4912]: I1203 00:23:41.623196 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:41 crc kubenswrapper[4912]: I1203 00:23:41.623214 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:43 crc kubenswrapper[4912]: I1203 00:23:43.395224 4912 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 00:23:43 crc kubenswrapper[4912]: I1203 00:23:43.395334 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:23:43 crc kubenswrapper[4912]: I1203 00:23:43.795983 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 00:23:43 crc kubenswrapper[4912]: I1203 00:23:43.796185 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:43 crc kubenswrapper[4912]: I1203 00:23:43.797299 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:43 crc kubenswrapper[4912]: I1203 00:23:43.797350 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:43 crc kubenswrapper[4912]: I1203 00:23:43.797360 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:44 crc kubenswrapper[4912]: E1203 00:23:44.638639 4912 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.219279 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.220361 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.221461 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.221509 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.221522 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.467100 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.467231 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.468219 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.468247 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.468259 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.471046 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.632841 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.634255 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.634299 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:46 crc kubenswrapper[4912]: I1203 00:23:46.634315 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:47 crc kubenswrapper[4912]: I1203 00:23:47.448189 4912 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 00:23:47 crc kubenswrapper[4912]: I1203 00:23:47.448255 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 00:23:47 crc kubenswrapper[4912]: I1203 00:23:47.451619 4912 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 00:23:47 crc kubenswrapper[4912]: I1203 00:23:47.451698 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.487983 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.488122 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.489116 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.489143 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.489154 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.491400 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.639090 4912 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.639143 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.640214 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.640285 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:49 crc kubenswrapper[4912]: I1203 00:23:49.640305 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.452755 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.455412 4912 trace.go:236] Trace[854291828]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 00:23:38.530) (total time: 13924ms): Dec 03 00:23:52 crc kubenswrapper[4912]: Trace[854291828]: ---"Objects listed" error: 13924ms (00:23:52.455) Dec 03 00:23:52 crc kubenswrapper[4912]: Trace[854291828]: [13.924399601s] [13.924399601s] END Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.455470 4912 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.456032 4912 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.456050 4912 trace.go:236] Trace[1695838339]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 00:23:38.219) (total time: 14236ms): Dec 03 00:23:52 crc kubenswrapper[4912]: Trace[1695838339]: ---"Objects listed" error: 14236ms (00:23:52.455) Dec 03 00:23:52 crc kubenswrapper[4912]: Trace[1695838339]: [14.236810394s] [14.236810394s] END Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.456079 4912 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.456370 4912 trace.go:236] Trace[1915286513]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 00:23:37.979) (total time: 14477ms): Dec 03 00:23:52 crc kubenswrapper[4912]: Trace[1915286513]: ---"Objects listed" error: 14477ms (00:23:52.456) Dec 03 00:23:52 crc kubenswrapper[4912]: Trace[1915286513]: [14.477087354s] [14.477087354s] END Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.456386 4912 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.456941 4912 trace.go:236] Trace[640471647]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 00:23:38.046) (total time: 14410ms): Dec 03 00:23:52 crc kubenswrapper[4912]: Trace[640471647]: ---"Objects listed" error: 14410ms (00:23:52.456) Dec 03 00:23:52 crc kubenswrapper[4912]: Trace[640471647]: [14.410616266s] [14.410616266s] END Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.456977 4912 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.458136 4912 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.475959 4912 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51634->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.476135 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51634->192.168.126.11:17697: read: connection reset by peer" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.476632 4912 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.476701 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.528410 4912 apiserver.go:52] "Watching apiserver" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.530596 4912 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.530882 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.531187 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.531210 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.531238 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.531666 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.531720 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.531829 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.531912 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.531959 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.531963 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.532821 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.533444 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.533539 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.534033 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.534066 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.534089 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.534926 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.535013 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.535319 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.554067 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.572152 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.582588 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.591064 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.599772 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.610808 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.622203 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.629627 4912 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.633050 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.646381 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.648521 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9" exitCode=255 Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.648564 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9"} Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657417 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657465 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657484 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657503 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657521 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657536 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657552 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657576 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657595 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657611 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657628 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657645 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657661 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657675 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657691 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657708 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657724 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657740 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657757 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657772 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657788 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657804 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657820 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657833 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657849 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657869 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657884 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657899 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657915 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657933 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657949 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657964 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657980 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657967 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.657994 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658039 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658055 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658090 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658106 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658121 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658137 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658210 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658210 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658229 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658246 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658280 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658302 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658314 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658321 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658418 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658458 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658508 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658538 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658564 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658588 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658615 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658622 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658641 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658668 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658691 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658697 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658721 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658751 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658777 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658796 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658801 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658830 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658843 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658846 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658843 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658873 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658897 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658920 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658945 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658969 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.658992 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659020 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659041 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659051 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659063 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659110 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659142 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659170 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659204 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659229 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659251 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659278 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659303 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659329 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659353 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659380 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659405 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659450 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659478 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659509 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659535 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659557 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659582 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659610 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659628 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659644 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659661 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659679 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659700 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659718 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659738 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659755 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659771 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659789 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659807 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659824 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659839 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659854 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659869 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659885 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659900 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659917 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659932 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659948 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659965 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659981 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660013 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660075 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660101 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660125 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660148 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660169 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660192 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660215 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660240 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660264 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660562 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660590 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660608 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660625 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660641 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660658 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660751 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660769 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659070 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659112 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659138 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659178 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659303 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659410 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659469 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659491 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.669901 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659641 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659669 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659767 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659920 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.659935 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660016 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.660874 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:23:53.160854036 +0000 UTC m=+18.802874596 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.660925 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.661063 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.661084 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.661218 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.661457 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.661456 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.661519 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.661654 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.661743 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.662185 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.662197 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.662224 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.662743 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.663047 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.663182 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.663210 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.663391 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.663347 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.663538 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.663576 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.663734 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.663674 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.664367 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.664388 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670219 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670151 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.664950 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670315 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.664973 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670345 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.664998 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.665566 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670363 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670454 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.665653 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.665919 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666021 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666183 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666295 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666400 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666427 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666533 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666554 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666780 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666791 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.666898 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667074 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667130 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667217 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667316 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667323 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667480 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667527 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667730 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667741 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667783 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.667833 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.668080 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.668372 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.668519 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.668540 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.668498 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.668550 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.668559 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.669558 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.669733 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.669787 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670058 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670127 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670180 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670906 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670396 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.664421 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670583 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670791 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670836 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.670982 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671090 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671138 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671333 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671365 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671384 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671419 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671485 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671499 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671514 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671545 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671692 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671726 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672202 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672232 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672258 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671561 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671953 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671922 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671966 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.671993 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672103 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672155 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672370 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672689 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672788 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672865 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672281 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672932 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673113 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673172 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673247 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673272 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673296 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673320 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673342 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673416 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673461 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673488 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673511 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673539 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673564 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673592 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673619 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673644 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673670 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673695 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673718 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673782 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.672893 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673168 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673202 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673998 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674006 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673413 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673427 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673672 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673890 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673896 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674393 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674460 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.673808 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674567 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674599 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674626 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674655 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674692 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674721 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674745 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674768 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674788 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674809 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674828 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674847 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.674864 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675079 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675081 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675096 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675115 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675133 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675134 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675150 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675169 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675188 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675190 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675205 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675221 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675237 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675252 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675269 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675287 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675304 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675320 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675336 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675356 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675379 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675914 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.676012 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.676118 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.676971 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.677747 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.678132 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.680151 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.680788 4912 scope.go:117] "RemoveContainer" containerID="e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682054 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682127 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682163 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682204 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682753 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682791 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682822 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682858 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682887 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682918 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682944 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683038 4912 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683058 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683072 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683088 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683103 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683129 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683143 4912 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683156 4912 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683172 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683187 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683200 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683212 4912 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683229 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683242 4912 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683256 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683269 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683287 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683300 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683312 4912 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683324 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683341 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683354 4912 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683366 4912 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683381 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683396 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683410 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683426 4912 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683495 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683508 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683521 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683533 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683549 4912 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683562 4912 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683574 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683586 4912 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683602 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683615 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683627 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683643 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683656 4912 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683669 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683682 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683698 4912 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683709 4912 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683720 4912 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683732 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683749 4912 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683761 4912 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683773 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683785 4912 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683802 4912 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683827 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683841 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683858 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683870 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683881 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675561 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675705 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.675985 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.676022 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.676864 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.677068 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.677358 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.677544 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.677911 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.678573 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.678645 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.678645 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.678727 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.678959 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.679022 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.679335 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.678041 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.679844 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.679868 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.680307 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.680231 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.680543 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.681018 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.684165 4912 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.684179 4912 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.684265 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:53.184250002 +0000 UTC m=+18.826270562 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.681294 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.681632 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.681051 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.681891 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682504 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682673 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682717 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682766 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682740 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682857 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682886 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.682907 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683054 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683105 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683212 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683289 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683377 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.683448 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.684525 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.684769 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.684923 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.684985 4912 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.685026 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:53.185013922 +0000 UTC m=+18.827034482 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685057 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685348 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685372 4912 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685384 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685394 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685404 4912 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685417 4912 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685465 4912 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685586 4912 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685667 4912 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685679 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685689 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685701 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685714 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685722 4912 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685732 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685740 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685752 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685762 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685771 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685783 4912 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685793 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685802 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685811 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685824 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685745 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685834 4912 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685875 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685899 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685916 4912 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.685929 4912 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686038 4912 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686061 4912 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686076 4912 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686090 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686103 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686101 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686122 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686121 4912 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686155 4912 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686171 4912 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686187 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686200 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686218 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686231 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686246 4912 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686258 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686274 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686286 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686299 4912 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686313 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686323 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686334 4912 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686345 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686359 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686372 4912 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686383 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686394 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686412 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686424 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686493 4912 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686509 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686521 4912 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686533 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686545 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686559 4912 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686570 4912 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686582 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686595 4912 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686610 4912 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686621 4912 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686633 4912 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686643 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686654 4912 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686663 4912 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686672 4912 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686683 4912 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.686692 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.687506 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.691672 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.697494 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.697625 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.697660 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.697688 4912 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.697727 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.697762 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:53.197733772 +0000 UTC m=+18.839754332 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.697831 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.697938 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.698150 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.698258 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.698354 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.698421 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.698500 4912 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:52 crc kubenswrapper[4912]: E1203 00:23:52.698610 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:53.198585235 +0000 UTC m=+18.840605795 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.698719 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.698774 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.700387 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.700506 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.701086 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.702324 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.703214 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.703216 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.706363 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.708372 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.709243 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.709361 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.709611 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.714087 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.714276 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.720314 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.730355 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.730853 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.736585 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787113 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787178 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787226 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787240 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787252 4912 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787263 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787273 4912 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787285 4912 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787296 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787308 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787319 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787342 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787356 4912 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787359 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787417 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787368 4912 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787468 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787480 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787490 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787499 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787509 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787518 4912 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787527 4912 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787536 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787545 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787555 4912 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787565 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787573 4912 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787584 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787596 4912 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787610 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787621 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787631 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787641 4912 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787651 4912 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787661 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787670 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787681 4912 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787690 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787699 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787708 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787718 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787728 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787743 4912 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787752 4912 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787761 4912 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787771 4912 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787779 4912 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787790 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787799 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787807 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787816 4912 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787825 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787834 4912 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787842 4912 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787851 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787859 4912 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787867 4912 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787876 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787885 4912 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787894 4912 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787902 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787911 4912 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787920 4912 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787929 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787936 4912 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.787944 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.845310 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.852644 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 00:23:52 crc kubenswrapper[4912]: W1203 00:23:52.858352 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-9dcb6e4aa7947a07460fb7d6c638a2ceb242c1171e6780c61d3527626467ddfa WatchSource:0}: Error finding container 9dcb6e4aa7947a07460fb7d6c638a2ceb242c1171e6780c61d3527626467ddfa: Status 404 returned error can't find the container with id 9dcb6e4aa7947a07460fb7d6c638a2ceb242c1171e6780c61d3527626467ddfa Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.859639 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 00:23:52 crc kubenswrapper[4912]: W1203 00:23:52.863386 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-51bb548105ca1230c721dba5e0808f07266efd5d0e299b3b6ba8bf616451caa6 WatchSource:0}: Error finding container 51bb548105ca1230c721dba5e0808f07266efd5d0e299b3b6ba8bf616451caa6: Status 404 returned error can't find the container with id 51bb548105ca1230c721dba5e0808f07266efd5d0e299b3b6ba8bf616451caa6 Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.875681 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.880016 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.882984 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.887723 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.897187 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.907216 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.923024 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.947540 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.957222 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.968980 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.979476 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:52 crc kubenswrapper[4912]: I1203 00:23:52.991484 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.007249 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.019151 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.030834 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.043833 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.053450 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.068687 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.191899 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.192003 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.192036 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.192119 4912 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.192126 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:23:54.192091325 +0000 UTC m=+19.834111885 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.192156 4912 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.192166 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:54.192152828 +0000 UTC m=+19.834173378 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.192265 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:54.1922437 +0000 UTC m=+19.834264330 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.293012 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.293080 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.293166 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.293166 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.293181 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.293188 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.293195 4912 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.293197 4912 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.293242 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:54.2932281 +0000 UTC m=+19.935248660 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.293255 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:54.293250171 +0000 UTC m=+19.935270721 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.614282 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.651483 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1ea135a8bafd89004410dcb9ddc682f4b8bdc81148ee3d33503675172d137880"} Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.653034 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49"} Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.653105 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393"} Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.653120 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"51bb548105ca1230c721dba5e0808f07266efd5d0e299b3b6ba8bf616451caa6"} Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.654104 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20"} Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.654130 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9dcb6e4aa7947a07460fb7d6c638a2ceb242c1171e6780c61d3527626467ddfa"} Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.655881 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.657540 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9"} Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.675943 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.696418 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.714708 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.737237 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.763830 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.777373 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.795461 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.821666 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.833774 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.836560 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.853208 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.856566 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.871477 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.905413 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.940318 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.978560 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.981110 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.983786 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-tnw7s"] Dec 03 00:23:53 crc kubenswrapper[4912]: I1203 00:23:53.984130 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tnw7s" Dec 03 00:23:53 crc kubenswrapper[4912]: W1203 00:23:53.989113 4912 reflector.go:561] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": failed to list *v1.Secret: secrets "node-resolver-dockercfg-kz9s7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.989171 4912 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"node-resolver-dockercfg-kz9s7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"node-resolver-dockercfg-kz9s7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 00:23:53 crc kubenswrapper[4912]: W1203 00:23:53.990454 4912 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.990504 4912 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 00:23:53 crc kubenswrapper[4912]: W1203 00:23:53.992988 4912 reflector.go:561] object-"openshift-dns"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Dec 03 00:23:53 crc kubenswrapper[4912]: E1203 00:23:53.993029 4912 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.015765 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.029500 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.043335 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.055850 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.068262 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.084175 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.096164 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.098704 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c92a06d2-780e-4372-9395-8519acdb616d-hosts-file\") pod \"node-resolver-tnw7s\" (UID: \"c92a06d2-780e-4372-9395-8519acdb616d\") " pod="openshift-dns/node-resolver-tnw7s" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.098758 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkkqq\" (UniqueName: \"kubernetes.io/projected/c92a06d2-780e-4372-9395-8519acdb616d-kube-api-access-hkkqq\") pod \"node-resolver-tnw7s\" (UID: \"c92a06d2-780e-4372-9395-8519acdb616d\") " pod="openshift-dns/node-resolver-tnw7s" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.114257 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.125686 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.137627 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.148788 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.158058 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.200125 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.200207 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.200235 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkkqq\" (UniqueName: \"kubernetes.io/projected/c92a06d2-780e-4372-9395-8519acdb616d-kube-api-access-hkkqq\") pod \"node-resolver-tnw7s\" (UID: \"c92a06d2-780e-4372-9395-8519acdb616d\") " pod="openshift-dns/node-resolver-tnw7s" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.200273 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c92a06d2-780e-4372-9395-8519acdb616d-hosts-file\") pod \"node-resolver-tnw7s\" (UID: \"c92a06d2-780e-4372-9395-8519acdb616d\") " pod="openshift-dns/node-resolver-tnw7s" Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.200326 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:23:56.20029551 +0000 UTC m=+21.842316070 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.200379 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c92a06d2-780e-4372-9395-8519acdb616d-hosts-file\") pod \"node-resolver-tnw7s\" (UID: \"c92a06d2-780e-4372-9395-8519acdb616d\") " pod="openshift-dns/node-resolver-tnw7s" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.200298 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.200416 4912 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.200535 4912 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.200543 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:56.200525926 +0000 UTC m=+21.842546486 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.200617 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:56.200599988 +0000 UTC m=+21.842620578 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.301830 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.301885 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.301983 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.302002 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.301985 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.302014 4912 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.302026 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.302035 4912 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.302064 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:56.3020481 +0000 UTC m=+21.944068660 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.302084 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:23:56.302075911 +0000 UTC m=+21.944096481 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.571532 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.571587 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.571618 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.571669 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.571754 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.571819 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.575050 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.575634 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.576909 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.577562 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.578523 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.579044 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.579636 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.580563 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.581128 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.582009 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.582605 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.583688 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.583813 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.584208 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.584711 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.585565 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.586041 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.587035 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.587387 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.587924 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.588832 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.589241 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.590159 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.590681 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.591724 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.592120 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.592747 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.593842 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.594277 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.595178 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.595751 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.596630 4912 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.596726 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.596715 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.598256 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.599104 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.599540 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.601085 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.601750 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.602693 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.603320 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.604307 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.604779 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.605816 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.606484 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.606572 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.607528 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.608010 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.611416 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.612225 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.613325 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.613838 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.614874 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.615423 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.616489 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.617021 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.617584 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.624768 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.635889 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.649924 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.660096 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.662424 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.668410 4912 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.674611 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.690030 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.704560 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.839654 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-2qk89"] Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.840060 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.840389 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-m2wz8"] Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.840756 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.848003 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.849231 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.850577 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.850807 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.850969 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.858635 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.858779 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.859207 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.859407 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.859746 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.864874 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f29xt"] Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.865750 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.875690 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-59mch"] Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.877081 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.888230 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 00:23:54 crc kubenswrapper[4912]: W1203 00:23:54.888251 4912 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": failed to list *v1.Secret: secrets "ovn-node-metrics-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 03 00:23:54 crc kubenswrapper[4912]: E1203 00:23:54.888547 4912 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-node-metrics-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.891670 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.891735 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.891865 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.891744 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.892134 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.895838 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.896096 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908008 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-conf-dir\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908062 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-var-lib-cni-bin\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908087 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-var-lib-kubelet\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908107 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-cni-dir\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908136 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/92796847-a53f-4ffd-8746-7fd65c0ee4ea-proxy-tls\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908155 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbpvw\" (UniqueName: \"kubernetes.io/projected/92796847-a53f-4ffd-8746-7fd65c0ee4ea-kube-api-access-lbpvw\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908176 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-run-k8s-cni-cncf-io\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908191 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-run-netns\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908210 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8pj4\" (UniqueName: \"kubernetes.io/projected/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-kube-api-access-m8pj4\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908229 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-cnibin\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908248 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/92796847-a53f-4ffd-8746-7fd65c0ee4ea-mcd-auth-proxy-config\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908262 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-os-release\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908279 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-var-lib-cni-multus\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908297 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-run-multus-certs\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908322 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-system-cni-dir\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908338 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-daemon-config\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908361 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-etc-kubernetes\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908451 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-socket-dir-parent\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908609 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/92796847-a53f-4ffd-8746-7fd65c0ee4ea-rootfs\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908645 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-cni-binary-copy\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.908662 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-hostroot\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:54 crc kubenswrapper[4912]: I1203 00:23:54.952755 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.008417 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009655 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-daemon-config\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009707 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-etc-openvswitch\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009733 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovn-node-metrics-cert\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009756 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-bin\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009797 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-hostroot\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009828 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/36b8f8fb-72ad-407e-914c-9f71ada6e67d-cni-binary-copy\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009848 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-systemd-units\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009861 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-slash\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009859 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-hostroot\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009882 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-var-lib-kubelet\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009930 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-var-lib-kubelet\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009933 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-os-release\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.009991 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-openvswitch\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010015 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8pj4\" (UniqueName: \"kubernetes.io/projected/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-kube-api-access-m8pj4\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010031 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-cnibin\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010046 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010062 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/92796847-a53f-4ffd-8746-7fd65c0ee4ea-proxy-tls\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010078 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010095 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-netd\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010112 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/92796847-a53f-4ffd-8746-7fd65c0ee4ea-mcd-auth-proxy-config\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010128 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-var-lib-cni-multus\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010146 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-run-multus-certs\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010160 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-kubelet\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010185 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-system-cni-dir\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010200 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-node-log\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010214 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-script-lib\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010230 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-etc-kubernetes\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010253 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-var-lib-openvswitch\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010267 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-log-socket\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010287 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs5bf\" (UniqueName: \"kubernetes.io/projected/7d5e7779-476a-4b6f-8153-7fd40fab8d90-kube-api-access-hs5bf\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010304 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/92796847-a53f-4ffd-8746-7fd65c0ee4ea-rootfs\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010318 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-cni-binary-copy\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010333 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-socket-dir-parent\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010350 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-var-lib-cni-bin\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010364 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-conf-dir\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010381 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-systemd\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010396 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-ovn-kubernetes\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010418 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-cni-dir\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010451 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-system-cni-dir\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010468 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-netns\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010484 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbpvw\" (UniqueName: \"kubernetes.io/projected/92796847-a53f-4ffd-8746-7fd65c0ee4ea-kube-api-access-lbpvw\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010500 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-run-k8s-cni-cncf-io\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010514 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-run-netns\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010529 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdrqq\" (UniqueName: \"kubernetes.io/projected/36b8f8fb-72ad-407e-914c-9f71ada6e67d-kube-api-access-cdrqq\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010548 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-ovn\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010566 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-env-overrides\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010587 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-cnibin\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010602 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/36b8f8fb-72ad-407e-914c-9f71ada6e67d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010617 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-os-release\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010620 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-daemon-config\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010631 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-config\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010892 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-socket-dir-parent\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010920 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-var-lib-cni-bin\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010941 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-conf-dir\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.010993 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-multus-cni-dir\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011126 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-cni-binary-copy\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011192 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-run-k8s-cni-cncf-io\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011231 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-run-netns\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011287 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-cnibin\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011314 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-system-cni-dir\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011537 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-etc-kubernetes\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011613 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/92796847-a53f-4ffd-8746-7fd65c0ee4ea-rootfs\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011655 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-os-release\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011688 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-run-multus-certs\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011711 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-host-var-lib-cni-multus\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.011914 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/92796847-a53f-4ffd-8746-7fd65c0ee4ea-mcd-auth-proxy-config\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.021393 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/92796847-a53f-4ffd-8746-7fd65c0ee4ea-proxy-tls\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.029574 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8pj4\" (UniqueName: \"kubernetes.io/projected/8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77-kube-api-access-m8pj4\") pod \"multus-m2wz8\" (UID: \"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\") " pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.033515 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbpvw\" (UniqueName: \"kubernetes.io/projected/92796847-a53f-4ffd-8746-7fd65c0ee4ea-kube-api-access-lbpvw\") pod \"machine-config-daemon-2qk89\" (UID: \"92796847-a53f-4ffd-8746-7fd65c0ee4ea\") " pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.035674 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.056478 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.072169 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.095725 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111385 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdrqq\" (UniqueName: \"kubernetes.io/projected/36b8f8fb-72ad-407e-914c-9f71ada6e67d-kube-api-access-cdrqq\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111422 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-ovn\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111460 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-env-overrides\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111494 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/36b8f8fb-72ad-407e-914c-9f71ada6e67d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111516 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-config\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111536 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-etc-openvswitch\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111555 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovn-node-metrics-cert\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111581 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-bin\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111619 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/36b8f8fb-72ad-407e-914c-9f71ada6e67d-cni-binary-copy\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111639 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-systemd-units\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111660 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-slash\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111685 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-os-release\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111708 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-openvswitch\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111727 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-cnibin\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111749 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111769 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-netd\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111792 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111817 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-kubelet\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111849 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-script-lib\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111869 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-node-log\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111892 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-var-lib-openvswitch\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111914 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-log-socket\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111934 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs5bf\" (UniqueName: \"kubernetes.io/projected/7d5e7779-476a-4b6f-8153-7fd40fab8d90-kube-api-access-hs5bf\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111970 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-systemd\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.111992 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-ovn-kubernetes\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.112015 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-system-cni-dir\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.112036 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-netns\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.112108 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-netns\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.112156 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-etc-openvswitch\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.112265 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-bin\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.112294 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-config\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.112359 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-ovn\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.112811 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-env-overrides\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113001 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/36b8f8fb-72ad-407e-914c-9f71ada6e67d-cni-binary-copy\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113059 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-systemd-units\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113092 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-slash\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113147 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-os-release\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113177 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-openvswitch\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113208 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-cnibin\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113306 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/36b8f8fb-72ad-407e-914c-9f71ada6e67d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113559 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113607 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-netd\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113608 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113644 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-var-lib-openvswitch\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113691 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113723 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-kubelet\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.113850 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-log-socket\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.114051 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-systemd\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.114094 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-ovn-kubernetes\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.114124 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/36b8f8fb-72ad-407e-914c-9f71ada6e67d-system-cni-dir\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.114156 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-node-log\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.114258 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-script-lib\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.122918 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.135307 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs5bf\" (UniqueName: \"kubernetes.io/projected/7d5e7779-476a-4b6f-8153-7fd40fab8d90-kube-api-access-hs5bf\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.136041 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdrqq\" (UniqueName: \"kubernetes.io/projected/36b8f8fb-72ad-407e-914c-9f71ada6e67d-kube-api-access-cdrqq\") pod \"multus-additional-cni-plugins-59mch\" (UID: \"36b8f8fb-72ad-407e-914c-9f71ada6e67d\") " pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.141342 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.155153 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.155145 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: W1203 00:23:55.165546 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92796847_a53f_4ffd_8746_7fd65c0ee4ea.slice/crio-03edd1ae5b1733811ee4a3474e3b96363b01048c4942b8ba353fc2cb4cd2d09f WatchSource:0}: Error finding container 03edd1ae5b1733811ee4a3474e3b96363b01048c4942b8ba353fc2cb4cd2d09f: Status 404 returned error can't find the container with id 03edd1ae5b1733811ee4a3474e3b96363b01048c4942b8ba353fc2cb4cd2d09f Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.165635 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m2wz8" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.169529 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: W1203 00:23:55.177023 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8260bfbd_d3da_4ea9_b9ad_1ac726fe0a77.slice/crio-2d3f784091ac2e7f704c740c496164eff41972a4bedb964104975dcc18158891 WatchSource:0}: Error finding container 2d3f784091ac2e7f704c740c496164eff41972a4bedb964104975dcc18158891: Status 404 returned error can't find the container with id 2d3f784091ac2e7f704c740c496164eff41972a4bedb964104975dcc18158891 Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.182218 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.197024 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-59mch" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.197016 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.211807 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: E1203 00:23:55.213931 4912 projected.go:288] Couldn't get configMap openshift-dns/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.227786 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.235674 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.239962 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.253096 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.276934 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.294858 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.314095 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.349723 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.377566 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.427799 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.441720 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.467842 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.489989 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.538205 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 00:23:55 crc kubenswrapper[4912]: E1203 00:23:55.544653 4912 projected.go:194] Error preparing data for projected volume kube-api-access-hkkqq for pod openshift-dns/node-resolver-tnw7s: failed to sync configmap cache: timed out waiting for the condition Dec 03 00:23:55 crc kubenswrapper[4912]: E1203 00:23:55.544732 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c92a06d2-780e-4372-9395-8519acdb616d-kube-api-access-hkkqq podName:c92a06d2-780e-4372-9395-8519acdb616d nodeName:}" failed. No retries permitted until 2025-12-03 00:23:56.044713422 +0000 UTC m=+21.686733972 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hkkqq" (UniqueName: "kubernetes.io/projected/c92a06d2-780e-4372-9395-8519acdb616d-kube-api-access-hkkqq") pod "node-resolver-tnw7s" (UID: "c92a06d2-780e-4372-9395-8519acdb616d") : failed to sync configmap cache: timed out waiting for the condition Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.658319 4912 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.662346 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.662398 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.662407 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.662515 4912 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.665364 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerStarted","Data":"f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.665417 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerStarted","Data":"aac17a2f852b35e696726c66947c449ba3eb98bb5f7534bca453d4c28b1de44b"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.666847 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2wz8" event={"ID":"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77","Type":"ContainerStarted","Data":"d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.666871 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2wz8" event={"ID":"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77","Type":"ContainerStarted","Data":"2d3f784091ac2e7f704c740c496164eff41972a4bedb964104975dcc18158891"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.668181 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.670359 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.670376 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.670385 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"03edd1ae5b1733811ee4a3474e3b96363b01048c4942b8ba353fc2cb4cd2d09f"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.783392 4912 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.783680 4912 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.784758 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.784781 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.784790 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.784803 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.784813 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:55Z","lastTransitionTime":"2025-12-03T00:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.799564 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.830900 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: E1203 00:23:55.832281 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.836790 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.836818 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.836826 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.836840 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.836849 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:55Z","lastTransitionTime":"2025-12-03T00:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:55 crc kubenswrapper[4912]: E1203 00:23:55.854189 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.854500 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.858119 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.858143 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.858150 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.858162 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.858170 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:55Z","lastTransitionTime":"2025-12-03T00:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.890827 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: E1203 00:23:55.892326 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.899445 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.899485 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.899496 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.899511 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.899522 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:55Z","lastTransitionTime":"2025-12-03T00:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.908159 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: E1203 00:23:55.917113 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.920669 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.920705 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.920714 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.920727 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.920737 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:55Z","lastTransitionTime":"2025-12-03T00:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.925220 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.925514 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: E1203 00:23:55.936008 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: E1203 00:23:55.936153 4912 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.936170 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovn-node-metrics-cert\") pod \"ovnkube-node-f29xt\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.937922 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.937960 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.937973 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.937989 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.938001 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:55Z","lastTransitionTime":"2025-12-03T00:23:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.938379 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.956191 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.970577 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.985124 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:55 crc kubenswrapper[4912]: I1203 00:23:55.996113 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.007270 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.026232 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.037604 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.039887 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.039919 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.039928 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.039943 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.039952 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.049856 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.064276 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.076062 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.088381 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.107533 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.121417 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkkqq\" (UniqueName: \"kubernetes.io/projected/c92a06d2-780e-4372-9395-8519acdb616d-kube-api-access-hkkqq\") pod \"node-resolver-tnw7s\" (UID: \"c92a06d2-780e-4372-9395-8519acdb616d\") " pod="openshift-dns/node-resolver-tnw7s" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.122487 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.136186 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.136910 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkkqq\" (UniqueName: \"kubernetes.io/projected/c92a06d2-780e-4372-9395-8519acdb616d-kube-api-access-hkkqq\") pod \"node-resolver-tnw7s\" (UID: \"c92a06d2-780e-4372-9395-8519acdb616d\") " pod="openshift-dns/node-resolver-tnw7s" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.143799 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.143856 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.143866 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.143885 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.143896 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.157264 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.182633 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.222163 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.222324 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.222413 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:24:00.222368993 +0000 UTC m=+25.864389553 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.222453 4912 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.222515 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:00.222498406 +0000 UTC m=+25.864519146 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.222505 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.222661 4912 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.222714 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:00.222706002 +0000 UTC m=+25.864726562 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.229837 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.242643 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.248353 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.248389 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.248400 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.248415 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.248441 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.265516 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.306652 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.324060 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.324114 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.324294 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.324315 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.324328 4912 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.324385 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:00.32436735 +0000 UTC m=+25.966387910 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.324802 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.324819 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.324829 4912 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.324854 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:00.324846083 +0000 UTC m=+25.966866643 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.344238 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.350857 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.350916 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.350934 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.350959 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.350975 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.383553 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.396712 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tnw7s" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.454230 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.454260 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.454270 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.454283 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.454292 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.557813 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.557863 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.557878 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.557895 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.557906 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.571461 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.571527 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.571593 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.571744 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.571862 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:56 crc kubenswrapper[4912]: E1203 00:23:56.571929 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.594147 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-wnn4k"] Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.595246 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.597489 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.597674 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.597789 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.597802 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.609807 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.627611 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.652671 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.660824 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.660857 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.660868 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.660902 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.660912 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.667123 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.677678 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01" exitCode=0 Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.677752 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.677780 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"35a3a7e8002b02de59d575e09d842d32a17f0e9d464e195bcd8bff14e04d9d8f"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.682029 4912 generic.go:334] "Generic (PLEG): container finished" podID="36b8f8fb-72ad-407e-914c-9f71ada6e67d" containerID="f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac" exitCode=0 Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.682099 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerDied","Data":"f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.683941 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tnw7s" event={"ID":"c92a06d2-780e-4372-9395-8519acdb616d","Type":"ContainerStarted","Data":"234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.683986 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tnw7s" event={"ID":"c92a06d2-780e-4372-9395-8519acdb616d","Type":"ContainerStarted","Data":"6e834ee730182b6bf08745ecc8e833ea2120c9e3f68a654a035f2127149a7501"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.695631 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.714837 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.728786 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/114ee369-a48e-4c87-9c8b-9436a230f02e-host\") pod \"node-ca-wnn4k\" (UID: \"114ee369-a48e-4c87-9c8b-9436a230f02e\") " pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.728847 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6b2v\" (UniqueName: \"kubernetes.io/projected/114ee369-a48e-4c87-9c8b-9436a230f02e-kube-api-access-w6b2v\") pod \"node-ca-wnn4k\" (UID: \"114ee369-a48e-4c87-9c8b-9436a230f02e\") " pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.728928 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/114ee369-a48e-4c87-9c8b-9436a230f02e-serviceca\") pod \"node-ca-wnn4k\" (UID: \"114ee369-a48e-4c87-9c8b-9436a230f02e\") " pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.752622 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.763306 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.763343 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.763353 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.763368 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.763378 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.786822 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.830285 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/114ee369-a48e-4c87-9c8b-9436a230f02e-host\") pod \"node-ca-wnn4k\" (UID: \"114ee369-a48e-4c87-9c8b-9436a230f02e\") " pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.830356 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6b2v\" (UniqueName: \"kubernetes.io/projected/114ee369-a48e-4c87-9c8b-9436a230f02e-kube-api-access-w6b2v\") pod \"node-ca-wnn4k\" (UID: \"114ee369-a48e-4c87-9c8b-9436a230f02e\") " pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.830447 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/114ee369-a48e-4c87-9c8b-9436a230f02e-serviceca\") pod \"node-ca-wnn4k\" (UID: \"114ee369-a48e-4c87-9c8b-9436a230f02e\") " pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.831223 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/114ee369-a48e-4c87-9c8b-9436a230f02e-serviceca\") pod \"node-ca-wnn4k\" (UID: \"114ee369-a48e-4c87-9c8b-9436a230f02e\") " pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.831588 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/114ee369-a48e-4c87-9c8b-9436a230f02e-host\") pod \"node-ca-wnn4k\" (UID: \"114ee369-a48e-4c87-9c8b-9436a230f02e\") " pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.833144 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.848580 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6b2v\" (UniqueName: \"kubernetes.io/projected/114ee369-a48e-4c87-9c8b-9436a230f02e-kube-api-access-w6b2v\") pod \"node-ca-wnn4k\" (UID: \"114ee369-a48e-4c87-9c8b-9436a230f02e\") " pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.866723 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.866768 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.866778 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.866796 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.866808 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.891026 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.906180 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wnn4k" Dec 03 00:23:56 crc kubenswrapper[4912]: W1203 00:23:56.919141 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod114ee369_a48e_4c87_9c8b_9436a230f02e.slice/crio-99db40ce8177994c9dd808e637ef5eb851dc8f157545e0a2f8d200afba4cfe81 WatchSource:0}: Error finding container 99db40ce8177994c9dd808e637ef5eb851dc8f157545e0a2f8d200afba4cfe81: Status 404 returned error can't find the container with id 99db40ce8177994c9dd808e637ef5eb851dc8f157545e0a2f8d200afba4cfe81 Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.924848 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.961515 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:56Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.969168 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.969201 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.969243 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.969256 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:56 crc kubenswrapper[4912]: I1203 00:23:56.969265 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:56Z","lastTransitionTime":"2025-12-03T00:23:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.020672 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.049774 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.071478 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.071750 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.071767 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.071784 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.071794 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:57Z","lastTransitionTime":"2025-12-03T00:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.082265 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.125287 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.163298 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.174598 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.174638 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.174648 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.174663 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.174672 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:57Z","lastTransitionTime":"2025-12-03T00:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.201826 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.244241 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.279920 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.279971 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.279985 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.280005 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.280017 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:57Z","lastTransitionTime":"2025-12-03T00:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.298528 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.340987 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.382788 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.382692 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.382833 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.382922 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.382944 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.382955 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:57Z","lastTransitionTime":"2025-12-03T00:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.410937 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.449535 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.482611 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.485244 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.485279 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.485288 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.485307 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.485319 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:57Z","lastTransitionTime":"2025-12-03T00:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.526087 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.569747 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.589931 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.589979 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.589989 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.590007 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.590017 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:57Z","lastTransitionTime":"2025-12-03T00:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.606041 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.643395 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.681891 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.694389 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.694422 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.694447 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.694465 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.694476 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:57Z","lastTransitionTime":"2025-12-03T00:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.696472 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerStarted","Data":"03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.701730 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.701796 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.701809 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.704823 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wnn4k" event={"ID":"114ee369-a48e-4c87-9c8b-9436a230f02e","Type":"ContainerStarted","Data":"d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.704875 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wnn4k" event={"ID":"114ee369-a48e-4c87-9c8b-9436a230f02e","Type":"ContainerStarted","Data":"99db40ce8177994c9dd808e637ef5eb851dc8f157545e0a2f8d200afba4cfe81"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.734085 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.775338 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.796456 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.796495 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.796507 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.796523 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.796533 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:57Z","lastTransitionTime":"2025-12-03T00:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.805561 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.844480 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.887475 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.899318 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.899359 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.899367 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.899384 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.899396 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:57Z","lastTransitionTime":"2025-12-03T00:23:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.923189 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:57 crc kubenswrapper[4912]: I1203 00:23:57.965253 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.001707 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.001762 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.001777 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.001803 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.001823 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.004801 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.043200 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.102944 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.105146 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.105188 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.105199 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.105218 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.105232 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.124478 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.171708 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.203938 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.207306 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.207335 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.207344 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.207358 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.207367 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.245617 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.280894 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.310025 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.310069 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.310080 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.310106 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.310118 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.328961 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.362283 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.401319 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.411836 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.411878 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.411896 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.411913 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.411924 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.440472 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.482900 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.514594 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.514636 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.514647 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.514663 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.514674 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.523135 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.561095 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.571309 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.571347 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.571309 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:23:58 crc kubenswrapper[4912]: E1203 00:23:58.571446 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:23:58 crc kubenswrapper[4912]: E1203 00:23:58.571542 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:23:58 crc kubenswrapper[4912]: E1203 00:23:58.571605 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.602587 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.616939 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.616981 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.616990 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.617006 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.617015 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.641647 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.686860 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.711809 4912 generic.go:334] "Generic (PLEG): container finished" podID="36b8f8fb-72ad-407e-914c-9f71ada6e67d" containerID="03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403" exitCode=0 Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.711855 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerDied","Data":"03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.715125 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.715165 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.715179 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.719654 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.719687 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.719699 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.719714 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.719725 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.727312 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.763905 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.803182 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.821851 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.821878 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.821885 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.821898 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.821906 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.842507 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.881948 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.921792 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.924241 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.924271 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.924291 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.924307 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.924317 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:58Z","lastTransitionTime":"2025-12-03T00:23:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.961931 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:58 crc kubenswrapper[4912]: I1203 00:23:58.999630 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:58Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.027180 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.027220 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.027230 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.027248 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.027257 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.042876 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.081506 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.121875 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.129835 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.129875 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.129883 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.129897 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.129906 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.162781 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.209250 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.231709 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.231749 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.231757 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.231772 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.231782 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.241535 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.287150 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.324757 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.334189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.334217 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.334228 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.334244 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.334256 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.368675 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.403842 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.436681 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.436720 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.436729 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.436745 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.436757 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.442631 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.481375 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.539578 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.539621 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.539631 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.539646 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.539655 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.641549 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.641591 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.641599 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.641612 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.641622 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.720530 4912 generic.go:334] "Generic (PLEG): container finished" podID="36b8f8fb-72ad-407e-914c-9f71ada6e67d" containerID="a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b" exitCode=0 Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.720577 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerDied","Data":"a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.733142 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.744044 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.744076 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.744087 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.744100 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.744110 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.745708 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.754859 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.767487 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.780073 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.790750 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.801642 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.811948 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.841143 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.846356 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.846389 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.846399 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.846414 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.846422 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.890029 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.922559 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.948721 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.948755 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.948766 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.948781 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.948793 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:23:59Z","lastTransitionTime":"2025-12-03T00:23:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:23:59 crc kubenswrapper[4912]: I1203 00:23:59.970390 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:23:59Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.004567 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.043079 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.050873 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.050904 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.050913 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.050930 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.050939 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.085123 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.153419 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.153475 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.153484 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.153497 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.153507 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.255759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.255802 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.255814 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.255831 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.255845 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.270542 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.270650 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.270708 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:24:08.270685392 +0000 UTC m=+33.912705952 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.270734 4912 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.270780 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.270821 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:08.270813035 +0000 UTC m=+33.912833595 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.270870 4912 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.270899 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:08.270893467 +0000 UTC m=+33.912914027 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.358524 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.358559 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.358568 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.358582 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.358592 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.372184 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.372219 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.372335 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.372350 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.372360 4912 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.372402 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:08.372390012 +0000 UTC m=+34.014410562 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.372466 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.372500 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.372514 4912 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.372583 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:08.372562256 +0000 UTC m=+34.014582906 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.461105 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.461145 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.461154 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.461168 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.461177 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.563896 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.564331 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.564455 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.564557 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.564638 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.571405 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.571455 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.571464 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.571540 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.571653 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:00 crc kubenswrapper[4912]: E1203 00:24:00.571751 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.666880 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.666906 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.666915 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.666928 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.666938 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.725577 4912 generic.go:334] "Generic (PLEG): container finished" podID="36b8f8fb-72ad-407e-914c-9f71ada6e67d" containerID="e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3" exitCode=0 Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.725646 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerDied","Data":"e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.729587 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.748341 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.764905 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.769854 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.769891 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.769899 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.769912 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.769920 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.776323 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.787225 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.803603 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.815004 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.832036 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.842906 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.853642 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.863902 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.872033 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.872070 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.872080 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.872095 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.872105 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.875597 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.886071 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.897347 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.913762 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.925631 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:00Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.974179 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.974403 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.974411 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.974438 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:00 crc kubenswrapper[4912]: I1203 00:24:00.974449 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:00Z","lastTransitionTime":"2025-12-03T00:24:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.374868 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.375386 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.375395 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.375412 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.375422 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:01Z","lastTransitionTime":"2025-12-03T00:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.478715 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.478749 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.478759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.478774 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.478783 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:01Z","lastTransitionTime":"2025-12-03T00:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.580725 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.580803 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.580813 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.580831 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.580842 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:01Z","lastTransitionTime":"2025-12-03T00:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.683239 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.683284 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.683294 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.683308 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.683318 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:01Z","lastTransitionTime":"2025-12-03T00:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.736773 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerStarted","Data":"bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a"} Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.755384 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.766698 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.778482 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.793994 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.794053 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.794069 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.794091 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.794108 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:01Z","lastTransitionTime":"2025-12-03T00:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.799445 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.811980 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.823094 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.831797 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.840774 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.857199 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.869222 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.880343 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.893583 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.896032 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.896072 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.896082 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.896097 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.896107 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:01Z","lastTransitionTime":"2025-12-03T00:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.904391 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.918551 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.931934 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:01Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.998262 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.998301 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.998312 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.998329 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:01 crc kubenswrapper[4912]: I1203 00:24:01.998340 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:01Z","lastTransitionTime":"2025-12-03T00:24:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.100679 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.100715 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.100727 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.100744 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.100756 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:02Z","lastTransitionTime":"2025-12-03T00:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.203167 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.203216 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.203231 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.203254 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.203269 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:02Z","lastTransitionTime":"2025-12-03T00:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.305574 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.305658 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.305682 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.305713 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.305737 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:02Z","lastTransitionTime":"2025-12-03T00:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.408322 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.408357 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.408366 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.408409 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.408422 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:02Z","lastTransitionTime":"2025-12-03T00:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.513549 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.513906 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.513920 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.513938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.513950 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:02Z","lastTransitionTime":"2025-12-03T00:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.570760 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.570827 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:02 crc kubenswrapper[4912]: E1203 00:24:02.570888 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.570939 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:02 crc kubenswrapper[4912]: E1203 00:24:02.570995 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:02 crc kubenswrapper[4912]: E1203 00:24:02.571099 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.616359 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.616381 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.616391 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.616404 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.616413 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:02Z","lastTransitionTime":"2025-12-03T00:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.718935 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.718970 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.718980 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.718997 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.719008 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:02Z","lastTransitionTime":"2025-12-03T00:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.744186 4912 generic.go:334] "Generic (PLEG): container finished" podID="36b8f8fb-72ad-407e-914c-9f71ada6e67d" containerID="bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a" exitCode=0 Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.744253 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerDied","Data":"bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.750628 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.751025 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.771565 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.784945 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.787445 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.796061 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.804546 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.815913 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.820494 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.820530 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.820540 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.820555 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.820566 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:02Z","lastTransitionTime":"2025-12-03T00:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.827868 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.841388 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.862086 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.875249 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.891825 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.922302 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.922327 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.922334 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.922346 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.922354 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:02Z","lastTransitionTime":"2025-12-03T00:24:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.962697 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.978011 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.988665 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:02 crc kubenswrapper[4912]: I1203 00:24:02.998832 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:02Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.008353 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.018705 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.024875 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.024905 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.024916 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.024931 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.024942 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.030482 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.039824 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.052150 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.067129 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.088952 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.107066 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.121986 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.127411 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.127474 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.127487 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.127503 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.127516 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.134084 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.145026 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.156450 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.176270 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.188521 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.198318 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.208542 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.230384 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.230424 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.230453 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.230471 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.230483 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.333958 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.333995 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.334004 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.334018 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.334028 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.436395 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.436448 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.436458 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.436470 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.436478 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.538940 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.538974 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.538983 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.538997 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.539008 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.641672 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.641715 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.641726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.641740 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.641749 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.743706 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.743752 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.743761 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.743776 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.743785 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.755300 4912 generic.go:334] "Generic (PLEG): container finished" podID="36b8f8fb-72ad-407e-914c-9f71ada6e67d" containerID="f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f" exitCode=0 Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.755367 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerDied","Data":"f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.755404 4912 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.755758 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.772456 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.783054 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.787154 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.798797 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.812472 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.822997 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.840128 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.845458 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.845482 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.845489 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.845502 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.845511 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.860369 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.872445 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.882384 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.890357 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.900152 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.912757 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.923018 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.938711 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.948533 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.948563 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.948573 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.948587 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.948596 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:03Z","lastTransitionTime":"2025-12-03T00:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.954261 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.967788 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.980401 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:03 crc kubenswrapper[4912]: I1203 00:24:03.990587 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:03Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.006600 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.019745 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.030559 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.042712 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.051217 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.051255 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.051273 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.051289 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.051300 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.052671 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.061819 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.080588 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.093358 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.102831 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.114025 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.126291 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.136831 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.153531 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.153565 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.153574 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.153588 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.153598 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.256000 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.256040 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.256054 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.256069 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.256079 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.358506 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.358546 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.358556 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.358575 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.358584 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.461088 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.461126 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.461136 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.461150 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.461158 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.563649 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.563694 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.563703 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.563955 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.563982 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.570973 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.570973 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:04 crc kubenswrapper[4912]: E1203 00:24:04.571088 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.571211 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:04 crc kubenswrapper[4912]: E1203 00:24:04.571459 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:04 crc kubenswrapper[4912]: E1203 00:24:04.571352 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.583571 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.603523 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.617308 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.633282 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.649302 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.661354 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.665709 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.665733 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.665742 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.665755 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.665764 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.674861 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.689538 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.710520 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.721593 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.731258 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.746303 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.760832 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" event={"ID":"36b8f8fb-72ad-407e-914c-9f71ada6e67d","Type":"ContainerStarted","Data":"16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.760928 4912 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.763621 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.767817 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.767858 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.767867 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.767881 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.767890 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.776668 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.808829 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.833899 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.845868 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.855207 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.864359 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.870225 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.870248 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.870255 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.870267 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.870275 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.875275 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.893491 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.901464 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.913394 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.926589 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.946253 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.975923 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.977683 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.977725 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.977740 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.977759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.977772 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:04Z","lastTransitionTime":"2025-12-03T00:24:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:04 crc kubenswrapper[4912]: I1203 00:24:04.995811 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.008681 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:05Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.025327 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:05Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.038399 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:05Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.080262 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.080309 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.080318 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.080332 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.080340 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:05Z","lastTransitionTime":"2025-12-03T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.182594 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.182633 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.182641 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.182655 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.182664 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:05Z","lastTransitionTime":"2025-12-03T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.285077 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.285122 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.285133 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.285148 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.285158 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:05Z","lastTransitionTime":"2025-12-03T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.387585 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.387633 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.387644 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.387657 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.387666 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:05Z","lastTransitionTime":"2025-12-03T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.490474 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.490793 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.490804 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.490820 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.490830 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:05Z","lastTransitionTime":"2025-12-03T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.616955 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.616979 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.616987 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.617000 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.617009 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:05Z","lastTransitionTime":"2025-12-03T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.719218 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.719259 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.719269 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.719291 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.719300 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:05Z","lastTransitionTime":"2025-12-03T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.764554 4912 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.820938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.820977 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.820985 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.820998 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.821007 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:05Z","lastTransitionTime":"2025-12-03T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.923470 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.923515 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.923523 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.923538 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:05 crc kubenswrapper[4912]: I1203 00:24:05.923556 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:05Z","lastTransitionTime":"2025-12-03T00:24:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.026471 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.026564 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.026583 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.026616 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.026638 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.030934 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.030975 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.030991 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.031008 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.031021 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: E1203 00:24:06.055723 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.061419 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.061503 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.061512 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.061527 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.061541 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: E1203 00:24:06.075090 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.079947 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.080000 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.080012 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.080032 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.080048 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: E1203 00:24:06.094318 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.099287 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.099324 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.099331 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.099347 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.099357 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: E1203 00:24:06.111789 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.115132 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.115176 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.115187 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.115202 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.115211 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: E1203 00:24:06.127790 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: E1203 00:24:06.127952 4912 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.129386 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.129423 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.129451 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.129468 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.129480 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.231628 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.231671 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.231684 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.231701 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.231712 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.334540 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.334577 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.334586 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.334601 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.334610 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.437169 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.437215 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.437228 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.437247 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.437303 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.539250 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.539293 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.539313 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.539335 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.539349 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.570637 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.570663 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.570731 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:06 crc kubenswrapper[4912]: E1203 00:24:06.570829 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:06 crc kubenswrapper[4912]: E1203 00:24:06.570898 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:06 crc kubenswrapper[4912]: E1203 00:24:06.570941 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.641869 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.641929 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.641943 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.641965 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.641990 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.744657 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.744738 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.744757 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.744786 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.744810 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.769748 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/0.log" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.772579 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b" exitCode=1 Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.772654 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.773750 4912 scope.go:117] "RemoveContainer" containerID="fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.789585 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.811388 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.826129 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.839868 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.846721 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.846760 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.846768 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.846782 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.846792 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.864140 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"203 00:24:05.581490 6177 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581553 6177 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581609 6177 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581615 6177 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581690 6177 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.582237 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:24:05.582248 6177 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:24:05.582328 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:24:05.582489 6177 factory.go:656] Stopping watch factory\\\\nI1203 00:24:05.582529 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:24:05.582541 6177 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:24:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.884142 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.909160 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.926933 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.941356 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.949649 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.949715 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.949727 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.949751 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.949768 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:06Z","lastTransitionTime":"2025-12-03T00:24:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.956553 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.971152 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.984675 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:06 crc kubenswrapper[4912]: I1203 00:24:06.996284 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:06Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.009092 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.023947 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.052380 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.052473 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.052485 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.052498 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.052508 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.155158 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.155212 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.155223 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.155239 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.155252 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.257598 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.257654 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.257667 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.257687 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.257700 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.360329 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.360389 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.360402 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.360426 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.360470 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.463081 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.463148 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.463165 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.463183 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.463193 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.565322 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.565377 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.565390 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.565408 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.565762 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.668284 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.668316 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.668324 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.668338 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.668347 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.771284 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.771331 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.771341 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.771358 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.771369 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.777476 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/0.log" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.779846 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.779989 4912 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.795424 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"203 00:24:05.581490 6177 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581553 6177 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581609 6177 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581615 6177 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581690 6177 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.582237 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:24:05.582248 6177 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:24:05.582328 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:24:05.582489 6177 factory.go:656] Stopping watch factory\\\\nI1203 00:24:05.582529 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:24:05.582541 6177 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:24:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.807722 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.819762 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.833729 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.849234 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.861128 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.873701 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.873733 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.873741 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.873755 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.873779 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.882244 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.895634 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.906803 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.916494 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.928513 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.941974 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.954844 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.975576 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.975646 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.975657 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.975705 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.975718 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:07Z","lastTransitionTime":"2025-12-03T00:24:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.979568 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:07 crc kubenswrapper[4912]: I1203 00:24:07.992153 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.078297 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.078359 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.078368 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.078381 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.078406 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:08Z","lastTransitionTime":"2025-12-03T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.180721 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.180761 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.180771 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.180791 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.180811 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:08Z","lastTransitionTime":"2025-12-03T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.282634 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.282698 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.282710 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.282727 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.282738 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:08Z","lastTransitionTime":"2025-12-03T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.351064 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.351189 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.351227 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:24:24.351206812 +0000 UTC m=+49.993227372 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.351253 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.351297 4912 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.351330 4912 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.351335 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:24.351325415 +0000 UTC m=+49.993345975 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.351352 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:24.351346515 +0000 UTC m=+49.993367075 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.385128 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.385167 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.385175 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.385188 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.385196 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:08Z","lastTransitionTime":"2025-12-03T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.414652 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944"] Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.415535 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.417902 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.418117 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.428976 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.441306 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.451260 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.451661 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.451702 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.451808 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.451832 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.451845 4912 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.451885 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:24.451870333 +0000 UTC m=+50.093890893 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.451808 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.451907 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.451916 4912 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.451944 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:24.451936545 +0000 UTC m=+50.093957105 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.465199 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.476760 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.487895 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.487927 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.487938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.487973 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.487983 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:08Z","lastTransitionTime":"2025-12-03T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.492474 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.505704 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.516010 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.531938 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"203 00:24:05.581490 6177 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581553 6177 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581609 6177 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581615 6177 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581690 6177 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.582237 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:24:05.582248 6177 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:24:05.582328 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:24:05.582489 6177 factory.go:656] Stopping watch factory\\\\nI1203 00:24:05.582529 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:24:05.582541 6177 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:24:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.549359 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.552555 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd86r\" (UniqueName: \"kubernetes.io/projected/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-kube-api-access-zd86r\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.552607 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.552724 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.552754 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.563634 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.571217 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.571270 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.571414 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.571538 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.571666 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.571786 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.577546 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.587864 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.590400 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.591020 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.591038 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.591057 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.591071 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:08Z","lastTransitionTime":"2025-12-03T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.599018 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.610792 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.619863 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.653458 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.653502 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.653524 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd86r\" (UniqueName: \"kubernetes.io/projected/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-kube-api-access-zd86r\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.653550 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.654026 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.654202 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.662948 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.668398 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd86r\" (UniqueName: \"kubernetes.io/projected/693d6bb6-3b42-408e-a6e2-43c9ac5fcda0-kube-api-access-zd86r\") pod \"ovnkube-control-plane-749d76644c-ng944\" (UID: \"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.693737 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.693781 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.693790 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.693807 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.693816 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:08Z","lastTransitionTime":"2025-12-03T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.729253 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" Dec 03 00:24:08 crc kubenswrapper[4912]: W1203 00:24:08.740592 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod693d6bb6_3b42_408e_a6e2_43c9ac5fcda0.slice/crio-da40f14109cd0849ad484f864b77ccacf219dfc61e9c7c279b5ea0bd20e95757 WatchSource:0}: Error finding container da40f14109cd0849ad484f864b77ccacf219dfc61e9c7c279b5ea0bd20e95757: Status 404 returned error can't find the container with id da40f14109cd0849ad484f864b77ccacf219dfc61e9c7c279b5ea0bd20e95757 Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.790587 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/1.log" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.791033 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/0.log" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.796842 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.796908 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.796921 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.796937 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.796949 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:08Z","lastTransitionTime":"2025-12-03T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.802036 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02" exitCode=1 Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.802125 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.802184 4912 scope.go:117] "RemoveContainer" containerID="fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.802948 4912 scope.go:117] "RemoveContainer" containerID="0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02" Dec 03 00:24:08 crc kubenswrapper[4912]: E1203 00:24:08.803108 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.807016 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" event={"ID":"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0","Type":"ContainerStarted","Data":"da40f14109cd0849ad484f864b77ccacf219dfc61e9c7c279b5ea0bd20e95757"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.819001 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.837392 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.851867 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.863302 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.876549 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.888328 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.898668 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.901128 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.901143 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.901150 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.901162 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.901170 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:08Z","lastTransitionTime":"2025-12-03T00:24:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.909307 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.923812 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.935198 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.948357 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.961770 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.981298 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"203 00:24:05.581490 6177 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581553 6177 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581609 6177 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581615 6177 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581690 6177 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.582237 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:24:05.582248 6177 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:24:05.582328 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:24:05.582489 6177 factory.go:656] Stopping watch factory\\\\nI1203 00:24:05.582529 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:24:05.582541 6177 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:24:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:08 crc kubenswrapper[4912]: I1203 00:24:08.996361 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.004528 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.004567 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.004578 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.004592 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.004602 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.005138 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.010340 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.023959 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.039087 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.052411 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.068855 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.090039 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"203 00:24:05.581490 6177 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581553 6177 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581609 6177 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581615 6177 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581690 6177 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.582237 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:24:05.582248 6177 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:24:05.582328 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:24:05.582489 6177 factory.go:656] Stopping watch factory\\\\nI1203 00:24:05.582529 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:24:05.582541 6177 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:24:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.107809 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.107869 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.107885 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.107910 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.107928 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.109346 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.126018 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.138811 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.150677 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.164085 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.183990 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.196770 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.207700 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.210173 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.210211 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.210219 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.210234 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.210243 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.216577 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.229850 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.242403 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.254165 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.312089 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.312142 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.312151 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.312164 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.312173 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.414368 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.414458 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.414473 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.414488 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.414497 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.516820 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.516856 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.516865 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.516878 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.516888 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.524348 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-kdznz"] Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.524819 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:09 crc kubenswrapper[4912]: E1203 00:24:09.524883 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.543014 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.554856 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.563994 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.564610 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.564656 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbfnp\" (UniqueName: \"kubernetes.io/projected/e49d5e0a-f59c-4dc1-9140-4b255576cd72-kube-api-access-hbfnp\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.572287 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.581780 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.592939 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.605423 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.618783 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.618818 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.618826 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.618840 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.618858 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.619523 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.632188 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.649737 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.661439 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.664920 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.664981 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbfnp\" (UniqueName: \"kubernetes.io/projected/e49d5e0a-f59c-4dc1-9140-4b255576cd72-kube-api-access-hbfnp\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:09 crc kubenswrapper[4912]: E1203 00:24:09.665087 4912 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:09 crc kubenswrapper[4912]: E1203 00:24:09.665132 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs podName:e49d5e0a-f59c-4dc1-9140-4b255576cd72 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:10.165119361 +0000 UTC m=+35.807139921 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs") pod "network-metrics-daemon-kdznz" (UID: "e49d5e0a-f59c-4dc1-9140-4b255576cd72") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.679060 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"203 00:24:05.581490 6177 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581553 6177 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581609 6177 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581615 6177 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581690 6177 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.582237 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:24:05.582248 6177 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:24:05.582328 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:24:05.582489 6177 factory.go:656] Stopping watch factory\\\\nI1203 00:24:05.582529 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:24:05.582541 6177 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:24:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.681116 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbfnp\" (UniqueName: \"kubernetes.io/projected/e49d5e0a-f59c-4dc1-9140-4b255576cd72-kube-api-access-hbfnp\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.695129 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.707756 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.721419 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.721479 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.721492 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.721509 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.721519 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.722781 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.733788 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.744906 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.811757 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" event={"ID":"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0","Type":"ContainerStarted","Data":"e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.811800 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" event={"ID":"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0","Type":"ContainerStarted","Data":"a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.813457 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/1.log" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.823000 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.823041 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.823050 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.823067 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.823080 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.827862 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.839516 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.856960 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.870125 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.880513 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.891132 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.901844 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.910968 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.921258 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.925137 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.925183 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.925195 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.925213 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.925224 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:09Z","lastTransitionTime":"2025-12-03T00:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.934317 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.946526 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.959117 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.971531 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:09 crc kubenswrapper[4912]: I1203 00:24:09.988186 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"203 00:24:05.581490 6177 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581553 6177 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581609 6177 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581615 6177 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581690 6177 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.582237 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:24:05.582248 6177 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:24:05.582328 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:24:05.582489 6177 factory.go:656] Stopping watch factory\\\\nI1203 00:24:05.582529 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:24:05.582541 6177 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:24:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:09Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.002564 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:10Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.015247 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:10Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.026705 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:10Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.026988 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.027021 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.027035 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.027050 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.027060 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.129303 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.129344 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.129354 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.129368 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.129379 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.168302 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:10 crc kubenswrapper[4912]: E1203 00:24:10.168511 4912 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:10 crc kubenswrapper[4912]: E1203 00:24:10.168599 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs podName:e49d5e0a-f59c-4dc1-9140-4b255576cd72 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:11.16857912 +0000 UTC m=+36.810599680 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs") pod "network-metrics-daemon-kdznz" (UID: "e49d5e0a-f59c-4dc1-9140-4b255576cd72") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.231396 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.231452 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.231464 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.231481 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.231494 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.333822 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.333865 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.333876 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.333891 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.333902 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.436112 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.436157 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.436167 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.436182 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.436191 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.538236 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.538296 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.538305 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.538318 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.538327 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.570873 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.570873 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.570890 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:10 crc kubenswrapper[4912]: E1203 00:24:10.571041 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:10 crc kubenswrapper[4912]: E1203 00:24:10.571123 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:10 crc kubenswrapper[4912]: E1203 00:24:10.571203 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.640938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.641013 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.641037 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.641060 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.641076 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.743929 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.743971 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.743984 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.744001 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.744012 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.846796 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.846837 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.846845 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.846858 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.846867 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.948937 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.948969 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.948982 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.948998 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:10 crc kubenswrapper[4912]: I1203 00:24:10.949010 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:10Z","lastTransitionTime":"2025-12-03T00:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.051309 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.051354 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.051364 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.051380 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.051390 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.153783 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.153821 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.153832 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.153844 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.153855 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.178575 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:11 crc kubenswrapper[4912]: E1203 00:24:11.178811 4912 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:11 crc kubenswrapper[4912]: E1203 00:24:11.178920 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs podName:e49d5e0a-f59c-4dc1-9140-4b255576cd72 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:13.178895232 +0000 UTC m=+38.820915862 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs") pod "network-metrics-daemon-kdznz" (UID: "e49d5e0a-f59c-4dc1-9140-4b255576cd72") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.256563 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.256612 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.256626 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.256643 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.256654 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.358605 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.358638 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.358646 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.358659 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.358669 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.460548 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.460593 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.460604 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.460618 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.460626 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.562727 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.562759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.562768 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.562782 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.562790 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.570995 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:11 crc kubenswrapper[4912]: E1203 00:24:11.571111 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.665027 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.665069 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.665079 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.665095 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.665106 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.767998 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.768033 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.768041 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.768054 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.768063 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.870323 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.870363 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.870372 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.870387 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.870399 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.972847 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.972890 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.972901 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.972917 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:11 crc kubenswrapper[4912]: I1203 00:24:11.972930 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:11Z","lastTransitionTime":"2025-12-03T00:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.074955 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.075011 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.075020 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.075035 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.075047 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.177288 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.177331 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.177342 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.177356 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.177366 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.279824 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.279858 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.279867 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.279879 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.279888 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.382714 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.382782 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.382794 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.382811 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.382824 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.485044 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.485087 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.485097 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.485115 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.485126 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.571201 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.571230 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:12 crc kubenswrapper[4912]: E1203 00:24:12.571331 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.571203 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:12 crc kubenswrapper[4912]: E1203 00:24:12.571516 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:12 crc kubenswrapper[4912]: E1203 00:24:12.571673 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.587010 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.587042 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.587051 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.587064 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.587073 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.689756 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.689797 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.689806 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.689828 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.689846 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.791764 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.791805 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.791817 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.791837 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.791851 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.894643 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.894682 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.894689 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.894704 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.894714 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.996829 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.996865 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.996873 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.996886 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:12 crc kubenswrapper[4912]: I1203 00:24:12.996894 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:12Z","lastTransitionTime":"2025-12-03T00:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.099119 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.099375 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.099504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.099603 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.099681 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:13Z","lastTransitionTime":"2025-12-03T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.198587 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:13 crc kubenswrapper[4912]: E1203 00:24:13.198693 4912 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:13 crc kubenswrapper[4912]: E1203 00:24:13.198737 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs podName:e49d5e0a-f59c-4dc1-9140-4b255576cd72 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:17.198723993 +0000 UTC m=+42.840744553 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs") pod "network-metrics-daemon-kdznz" (UID: "e49d5e0a-f59c-4dc1-9140-4b255576cd72") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.202804 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.202869 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.202887 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.202916 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.202938 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:13Z","lastTransitionTime":"2025-12-03T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.305400 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.305507 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.305520 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.305538 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.305573 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:13Z","lastTransitionTime":"2025-12-03T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.408078 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.408140 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.408152 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.408168 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.408179 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:13Z","lastTransitionTime":"2025-12-03T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.510341 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.510382 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.510392 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.510406 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.510416 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:13Z","lastTransitionTime":"2025-12-03T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.571318 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:13 crc kubenswrapper[4912]: E1203 00:24:13.571472 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.612552 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.612604 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.612615 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.612628 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.612637 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:13Z","lastTransitionTime":"2025-12-03T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.715227 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.715261 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.715270 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.715282 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.715291 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:13Z","lastTransitionTime":"2025-12-03T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.817794 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.817846 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.817854 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.817897 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.817909 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:13Z","lastTransitionTime":"2025-12-03T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.919751 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.919788 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.919798 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.919814 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:13 crc kubenswrapper[4912]: I1203 00:24:13.919825 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:13Z","lastTransitionTime":"2025-12-03T00:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.022647 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.022690 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.022700 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.022713 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.022722 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.124655 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.124703 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.124718 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.124738 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.124753 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.226855 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.226902 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.226912 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.226928 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.226937 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.329353 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.329395 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.329405 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.329418 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.329442 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.432250 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.432322 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.432347 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.432368 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.432383 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.536079 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.536116 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.536132 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.536149 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.536162 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.571501 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.571543 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.571506 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:14 crc kubenswrapper[4912]: E1203 00:24:14.571629 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:14 crc kubenswrapper[4912]: E1203 00:24:14.571728 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:14 crc kubenswrapper[4912]: E1203 00:24:14.571862 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.589389 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.610124 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fcd4ae6bd0f76e2b27c773bfe98d33256afba4c096b7887b1975bc4fa0198e6b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:06Z\\\",\\\"message\\\":\\\"203 00:24:05.581490 6177 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581553 6177 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 00:24:05.581609 6177 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581615 6177 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.581690 6177 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 00:24:05.582237 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 00:24:05.582248 6177 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 00:24:05.582328 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 00:24:05.582489 6177 factory.go:656] Stopping watch factory\\\\nI1203 00:24:05.582529 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1203 00:24:05.582541 6177 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1203 00:24:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.626413 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.638248 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.638509 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.638603 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.638730 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.638816 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.640037 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.659244 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.670107 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.682500 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.702065 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.713698 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.724601 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.735082 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.740913 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.740949 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.740957 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.740972 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.740984 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.746185 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.756803 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.764890 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.776769 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.788027 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.797068 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.842818 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.842854 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.842866 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.842881 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.842894 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.945300 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.945352 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.945369 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.945392 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:14 crc kubenswrapper[4912]: I1203 00:24:14.945406 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:14Z","lastTransitionTime":"2025-12-03T00:24:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.047731 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.047775 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.047790 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.047805 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.047815 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.150451 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.150494 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.150504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.150519 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.150527 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.253499 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.253546 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.253563 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.253583 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.253600 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.356232 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.356298 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.356362 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.356388 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.356405 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.459141 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.459187 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.459197 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.459212 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.459221 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.561702 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.561788 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.561809 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.561834 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.561853 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.571061 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:15 crc kubenswrapper[4912]: E1203 00:24:15.571233 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.664454 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.664499 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.664511 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.664526 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.664538 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.766805 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.766886 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.766910 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.766943 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.766966 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.869806 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.869870 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.869887 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.869911 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.869927 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.972185 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.972238 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.972622 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.972661 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:15 crc kubenswrapper[4912]: I1203 00:24:15.972678 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:15Z","lastTransitionTime":"2025-12-03T00:24:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.075717 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.075783 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.075798 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.075818 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.075832 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.143367 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.143420 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.143454 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.143473 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.143485 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: E1203 00:24:16.155858 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.159674 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.159718 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.159730 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.159748 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.159759 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: E1203 00:24:16.171517 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.175845 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.175909 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.175925 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.175946 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.175963 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: E1203 00:24:16.193830 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.198032 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.198078 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.198086 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.198112 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.198122 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: E1203 00:24:16.211899 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.216369 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.216410 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.216418 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.216454 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.216464 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: E1203 00:24:16.231002 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:16Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:16 crc kubenswrapper[4912]: E1203 00:24:16.231123 4912 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.232770 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.232803 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.232812 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.232828 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.232839 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.335070 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.335121 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.335143 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.335158 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.335169 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.437456 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.437495 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.437504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.437518 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.437529 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.539885 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.539922 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.539930 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.539945 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.539954 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.570805 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.570911 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.571018 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:16 crc kubenswrapper[4912]: E1203 00:24:16.571055 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:16 crc kubenswrapper[4912]: E1203 00:24:16.570929 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:16 crc kubenswrapper[4912]: E1203 00:24:16.571140 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.642255 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.642345 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.642355 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.642369 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.642379 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.744490 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.744532 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.744550 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.744566 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.744580 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.846724 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.846769 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.846779 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.846799 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.846810 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.948918 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.948989 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.948997 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.949012 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:16 crc kubenswrapper[4912]: I1203 00:24:16.949020 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:16Z","lastTransitionTime":"2025-12-03T00:24:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.051374 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.051759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.051772 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.051788 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.051799 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.155151 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.155189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.155197 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.155209 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.155218 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.238925 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:17 crc kubenswrapper[4912]: E1203 00:24:17.239069 4912 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:17 crc kubenswrapper[4912]: E1203 00:24:17.239150 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs podName:e49d5e0a-f59c-4dc1-9140-4b255576cd72 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:25.239132978 +0000 UTC m=+50.881153538 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs") pod "network-metrics-daemon-kdznz" (UID: "e49d5e0a-f59c-4dc1-9140-4b255576cd72") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.256920 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.256953 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.256964 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.256977 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.256988 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.360031 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.360098 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.360115 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.360137 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.360155 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.462919 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.462960 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.462970 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.462987 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.462998 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.564919 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.564957 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.564965 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.564977 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.564986 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.571142 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:17 crc kubenswrapper[4912]: E1203 00:24:17.571264 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.668116 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.668157 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.668166 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.668180 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.668191 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.770493 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.770527 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.770539 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.770553 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.770563 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.872915 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.872955 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.872964 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.872977 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.872985 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.975336 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.975371 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.975379 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.975392 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:17 crc kubenswrapper[4912]: I1203 00:24:17.975402 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:17Z","lastTransitionTime":"2025-12-03T00:24:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.077817 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.077861 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.077878 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.077892 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.077902 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.180071 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.180117 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.180128 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.180142 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.180153 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.282203 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.282242 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.282254 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.282270 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.282283 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.384620 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.384680 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.384697 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.384719 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.384737 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.487182 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.487251 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.487262 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.487275 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.487286 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.571487 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.571511 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.571617 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:18 crc kubenswrapper[4912]: E1203 00:24:18.571752 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:18 crc kubenswrapper[4912]: E1203 00:24:18.571844 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:18 crc kubenswrapper[4912]: E1203 00:24:18.571928 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.589576 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.589606 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.589614 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.589629 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.589638 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.692406 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.692514 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.692529 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.692547 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.692563 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.794700 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.794750 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.794765 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.794785 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.794800 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.896965 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.897004 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.897016 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.897035 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.897045 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.998820 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.998869 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.998882 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.998896 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:18 crc kubenswrapper[4912]: I1203 00:24:18.998906 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:18Z","lastTransitionTime":"2025-12-03T00:24:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.101705 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.101748 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.101759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.101772 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.101783 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:19Z","lastTransitionTime":"2025-12-03T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.204608 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.204653 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.204681 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.204696 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.204706 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:19Z","lastTransitionTime":"2025-12-03T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.307815 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.307866 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.307877 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.307896 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.307908 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:19Z","lastTransitionTime":"2025-12-03T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.409910 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.409955 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.409972 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.409992 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.410005 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:19Z","lastTransitionTime":"2025-12-03T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.511984 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.512023 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.512032 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.512048 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.512057 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:19Z","lastTransitionTime":"2025-12-03T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.570947 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:19 crc kubenswrapper[4912]: E1203 00:24:19.571073 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.615326 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.615370 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.615380 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.615397 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.615416 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:19Z","lastTransitionTime":"2025-12-03T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.717328 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.717365 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.717374 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.717390 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.717401 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:19Z","lastTransitionTime":"2025-12-03T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.819622 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.819670 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.819681 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.819700 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.819711 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:19Z","lastTransitionTime":"2025-12-03T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.921657 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.921699 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.921709 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.921727 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:19 crc kubenswrapper[4912]: I1203 00:24:19.921738 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:19Z","lastTransitionTime":"2025-12-03T00:24:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.024080 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.024123 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.024131 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.024148 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.024158 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.126006 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.126049 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.126063 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.126075 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.126083 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.227959 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.227997 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.228005 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.228019 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.228028 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.330171 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.330218 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.330229 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.330249 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.330260 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.432926 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.432971 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.432980 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.433000 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.433009 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.535977 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.536030 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.536039 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.536056 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.536066 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.571111 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.571183 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.571252 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:20 crc kubenswrapper[4912]: E1203 00:24:20.571338 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:20 crc kubenswrapper[4912]: E1203 00:24:20.571420 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:20 crc kubenswrapper[4912]: E1203 00:24:20.571588 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.638563 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.638595 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.638603 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.638616 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.638624 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.740756 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.740801 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.740814 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.740829 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.740839 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.843066 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.843101 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.843110 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.843124 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.843132 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.946476 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.946534 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.946546 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.946569 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:20 crc kubenswrapper[4912]: I1203 00:24:20.946582 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:20Z","lastTransitionTime":"2025-12-03T00:24:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.049349 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.049390 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.049399 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.049415 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.049424 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.150890 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.150931 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.150942 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.150956 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.150967 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.252974 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.253009 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.253022 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.253042 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.253058 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.355077 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.355115 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.355125 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.355139 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.355151 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.457626 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.457686 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.457703 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.457726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.457740 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.560656 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.560700 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.560712 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.560729 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.560739 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.570962 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:21 crc kubenswrapper[4912]: E1203 00:24:21.571135 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.663740 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.663824 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.663841 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.663868 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.663884 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.766184 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.766241 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.766251 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.766267 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.766280 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.869551 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.869723 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.870083 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.870135 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.870161 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.974814 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.974854 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.974864 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.974881 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:21 crc kubenswrapper[4912]: I1203 00:24:21.974898 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:21Z","lastTransitionTime":"2025-12-03T00:24:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.076933 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.076974 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.076983 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.076995 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.077004 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.179158 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.179202 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.179213 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.179229 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.179239 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.281238 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.281271 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.281280 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.281293 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.281302 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.383978 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.384036 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.384053 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.384074 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.384090 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.487112 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.487192 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.487215 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.487242 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.487273 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.571507 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.571711 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:22 crc kubenswrapper[4912]: E1203 00:24:22.571798 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.571818 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:22 crc kubenswrapper[4912]: E1203 00:24:22.571882 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:22 crc kubenswrapper[4912]: E1203 00:24:22.572585 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.589049 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.589169 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.589195 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.589222 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.589243 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.691375 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.691415 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.691440 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.691456 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.691467 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.793938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.793964 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.793974 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.793988 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.793997 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.896545 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.896586 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.896597 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.896611 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.896621 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.999221 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.999258 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.999269 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.999284 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:22 crc kubenswrapper[4912]: I1203 00:24:22.999295 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:22Z","lastTransitionTime":"2025-12-03T00:24:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.102417 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.102490 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.102504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.102540 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.102559 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:23Z","lastTransitionTime":"2025-12-03T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.204748 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.204789 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.204798 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.204812 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.204822 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:23Z","lastTransitionTime":"2025-12-03T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.307014 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.307048 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.307058 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.307071 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.307080 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:23Z","lastTransitionTime":"2025-12-03T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.409697 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.409747 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.409759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.409775 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.409785 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:23Z","lastTransitionTime":"2025-12-03T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.511943 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.511973 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.511982 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.511995 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.512005 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:23Z","lastTransitionTime":"2025-12-03T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.571140 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:23 crc kubenswrapper[4912]: E1203 00:24:23.571313 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.572125 4912 scope.go:117] "RemoveContainer" containerID="0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.586994 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.604695 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.614016 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.614043 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.614051 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.614127 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.614139 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:23Z","lastTransitionTime":"2025-12-03T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.617401 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.650760 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.668444 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.680016 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.692151 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.705551 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.716751 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.716783 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.716792 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.716805 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.716816 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:23Z","lastTransitionTime":"2025-12-03T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.718980 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.739170 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.751943 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.764961 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.776535 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.789558 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.803934 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.816320 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.818987 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.819216 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.819227 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.819241 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.819250 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:23Z","lastTransitionTime":"2025-12-03T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.826994 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.859048 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/1.log" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.861048 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.861177 4912 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.871518 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.884097 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.902949 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.920988 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.921015 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.921024 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.921036 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.921044 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:23Z","lastTransitionTime":"2025-12-03T00:24:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.922508 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.939981 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.958891 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:23 crc kubenswrapper[4912]: I1203 00:24:23.984516 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.004129 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:23Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.023792 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.023828 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.023836 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.023850 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.023859 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.026169 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.044798 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.061527 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.081204 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.093242 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.103725 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.114989 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.128336 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.129373 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.129402 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.129419 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.129447 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.129457 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.140670 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.231632 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.231680 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.231694 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.231710 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.231723 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.339754 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.339791 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.339802 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.339820 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.339831 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.415218 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.415323 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.415359 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.415382 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:24:56.415360504 +0000 UTC m=+82.057381064 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.415422 4912 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.415479 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:56.415470927 +0000 UTC m=+82.057491487 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.415556 4912 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.415652 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:56.415628982 +0000 UTC m=+82.057649542 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.442307 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.442366 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.442381 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.442397 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.442407 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.516624 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.516681 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.516790 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.516806 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.516832 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.516846 4912 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.516895 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:56.516880038 +0000 UTC m=+82.158900608 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.516809 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.516913 4912 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.516948 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:56.5169387 +0000 UTC m=+82.158959260 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.545376 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.545459 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.545471 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.545487 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.545498 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.570833 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.570876 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.570939 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.570994 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.571031 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:24 crc kubenswrapper[4912]: E1203 00:24:24.571120 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.584239 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.593369 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.605220 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.617659 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.629767 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.639221 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.646991 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.647010 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.647019 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.647032 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.647043 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.653671 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.665053 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.674520 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.689613 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.705986 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.717184 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.730782 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.741086 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.749154 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.749181 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.749188 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.749201 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.749209 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.750881 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.761135 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.778680 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.851150 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.851189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.851200 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.851216 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.851228 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.869335 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.954041 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.954128 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.954144 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.954168 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:24 crc kubenswrapper[4912]: I1203 00:24:24.954185 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:24Z","lastTransitionTime":"2025-12-03T00:24:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.057143 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.057191 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.057205 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.057226 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.057241 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.160015 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.160067 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.160082 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.160105 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.160120 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.263515 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.263576 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.263594 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.263616 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.263634 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.324738 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:25 crc kubenswrapper[4912]: E1203 00:24:25.324909 4912 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:25 crc kubenswrapper[4912]: E1203 00:24:25.325036 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs podName:e49d5e0a-f59c-4dc1-9140-4b255576cd72 nodeName:}" failed. No retries permitted until 2025-12-03 00:24:41.325003834 +0000 UTC m=+66.967024424 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs") pod "network-metrics-daemon-kdznz" (UID: "e49d5e0a-f59c-4dc1-9140-4b255576cd72") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.367246 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.367350 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.367362 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.367379 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.367389 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.470854 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.470926 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.470944 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.470970 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.470986 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.570770 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:25 crc kubenswrapper[4912]: E1203 00:24:25.570925 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.572977 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.573066 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.573083 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.573097 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.573107 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.675556 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.675596 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.675608 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.675624 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.675634 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.778479 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.778519 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.778531 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.778545 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.778556 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.869356 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/2.log" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.870481 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/1.log" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.874089 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63" exitCode=1 Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.874139 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.874188 4912 scope.go:117] "RemoveContainer" containerID="0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.876278 4912 scope.go:117] "RemoveContainer" containerID="e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63" Dec 03 00:24:25 crc kubenswrapper[4912]: E1203 00:24:25.876626 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.881141 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.881258 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.881357 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.881455 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.881539 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.889742 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.902940 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.915154 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.925651 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.936196 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.945495 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.961873 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.975596 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.983788 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.983873 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.983883 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.983897 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.983905 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:25Z","lastTransitionTime":"2025-12-03T00:24:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:25 crc kubenswrapper[4912]: I1203 00:24:25.988944 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.000822 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:25Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.010892 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.019082 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.035016 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.048671 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.056683 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.065849 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.075867 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.086148 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.086183 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.086192 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.086210 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.086221 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.188683 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.188725 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.188735 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.188751 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.188761 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.223408 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.233987 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.234309 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.246151 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.258172 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.271048 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.284694 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.291557 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.291770 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.291858 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.291938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.292087 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.297156 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.308300 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.317643 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.327989 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.344342 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.356579 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.366795 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.377470 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.385853 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.394542 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.394603 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.394615 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.394634 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.394646 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.394792 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.411125 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.422701 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.497575 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.497633 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.497648 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.497665 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.497677 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.571070 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:26 crc kubenswrapper[4912]: E1203 00:24:26.571188 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.571080 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:26 crc kubenswrapper[4912]: E1203 00:24:26.571525 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.571749 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:26 crc kubenswrapper[4912]: E1203 00:24:26.572065 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.593370 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.593681 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.593805 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.593932 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.594046 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: E1203 00:24:26.606110 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.609552 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.609591 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.609602 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.609615 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.609624 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: E1203 00:24:26.620717 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.624150 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.624192 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.624203 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.624221 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.624234 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: E1203 00:24:26.635081 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.638704 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.638784 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.638806 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.638832 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.638849 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: E1203 00:24:26.654130 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.658107 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.658162 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.658179 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.658205 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.658245 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: E1203 00:24:26.673118 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:26Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:26 crc kubenswrapper[4912]: E1203 00:24:26.673236 4912 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.674862 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.674889 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.674901 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.674916 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.674928 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.776834 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.776876 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.776940 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.776956 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.776965 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.878795 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/2.log" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.879260 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.879363 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.879382 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.879410 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.879452 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.981882 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.981917 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.981928 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.981943 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:26 crc kubenswrapper[4912]: I1203 00:24:26.981954 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:26Z","lastTransitionTime":"2025-12-03T00:24:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.084378 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.084423 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.084456 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.084477 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.084490 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:27Z","lastTransitionTime":"2025-12-03T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.187420 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.187480 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.187491 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.187506 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.187516 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:27Z","lastTransitionTime":"2025-12-03T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.289671 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.289720 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.289739 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.289764 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.289777 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:27Z","lastTransitionTime":"2025-12-03T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.392080 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.392112 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.392121 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.392134 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.392143 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:27Z","lastTransitionTime":"2025-12-03T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.494185 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.494228 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.494238 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.494254 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.494266 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:27Z","lastTransitionTime":"2025-12-03T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.571007 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:27 crc kubenswrapper[4912]: E1203 00:24:27.571121 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.596709 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.596761 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.596778 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.596812 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.596827 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:27Z","lastTransitionTime":"2025-12-03T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.700096 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.700149 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.700164 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.700185 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.700201 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:27Z","lastTransitionTime":"2025-12-03T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.802799 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.802834 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.802843 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.802855 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.802864 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:27Z","lastTransitionTime":"2025-12-03T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.904800 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.904840 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.904852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.904869 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:27 crc kubenswrapper[4912]: I1203 00:24:27.904882 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:27Z","lastTransitionTime":"2025-12-03T00:24:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.007030 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.007068 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.007084 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.007102 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.007114 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.110665 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.110714 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.110722 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.110736 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.110745 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.213365 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.213412 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.213423 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.213471 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.213483 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.316387 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.316542 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.316561 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.316622 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.316641 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.418741 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.418794 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.418805 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.418819 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.418829 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.520771 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.520813 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.520822 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.520834 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.520842 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.570772 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.570814 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.570781 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:28 crc kubenswrapper[4912]: E1203 00:24:28.570895 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:28 crc kubenswrapper[4912]: E1203 00:24:28.570970 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:28 crc kubenswrapper[4912]: E1203 00:24:28.571021 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.623536 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.623570 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.623578 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.623592 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.623602 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.725864 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.725917 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.725928 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.725945 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.725955 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.829054 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.829109 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.829121 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.829141 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.829155 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.932260 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.932298 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.932308 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.932324 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:28 crc kubenswrapper[4912]: I1203 00:24:28.932335 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:28Z","lastTransitionTime":"2025-12-03T00:24:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.034838 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.034872 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.034882 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.034895 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.034904 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.137337 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.137371 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.137380 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.137394 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.137402 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.239208 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.239297 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.239309 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.239322 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.239331 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.342247 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.342283 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.342292 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.342308 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.342325 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.444583 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.444626 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.444635 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.444649 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.444658 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.547114 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.547184 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.547194 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.547209 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.547218 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.570739 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:29 crc kubenswrapper[4912]: E1203 00:24:29.570861 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.649857 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.649921 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.649947 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.649977 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.650000 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.752787 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.752830 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.752842 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.752856 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.752865 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.856659 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.856723 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.856742 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.856763 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.856778 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.959794 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.959872 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.959884 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.959907 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:29 crc kubenswrapper[4912]: I1203 00:24:29.959923 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:29Z","lastTransitionTime":"2025-12-03T00:24:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.062358 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.062393 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.062422 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.062464 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.062474 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.165342 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.165409 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.165455 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.165505 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.165518 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.268911 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.268967 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.268976 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.268997 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.269007 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.371495 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.371545 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.371554 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.371567 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.371576 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.474958 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.475007 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.475017 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.475035 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.475045 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.571278 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.571477 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.571484 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:30 crc kubenswrapper[4912]: E1203 00:24:30.571645 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:30 crc kubenswrapper[4912]: E1203 00:24:30.571884 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:30 crc kubenswrapper[4912]: E1203 00:24:30.571981 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.577970 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.578007 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.578024 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.578045 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.578061 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.680137 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.680189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.680202 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.680218 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.680229 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.782982 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.783055 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.783083 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.783121 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.783147 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.886037 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.886126 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.886140 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.886167 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.886183 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.989180 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.989242 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.989251 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.989266 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:30 crc kubenswrapper[4912]: I1203 00:24:30.989280 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:30Z","lastTransitionTime":"2025-12-03T00:24:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.092309 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.092405 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.092418 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.092458 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.092470 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:31Z","lastTransitionTime":"2025-12-03T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.195132 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.195407 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.195424 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.195488 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.195519 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:31Z","lastTransitionTime":"2025-12-03T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.299057 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.299108 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.299123 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.299147 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.299165 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:31Z","lastTransitionTime":"2025-12-03T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.401784 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.401845 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.401857 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.401885 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.401900 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:31Z","lastTransitionTime":"2025-12-03T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.504762 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.504802 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.504812 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.504828 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.504840 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:31Z","lastTransitionTime":"2025-12-03T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.570916 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:31 crc kubenswrapper[4912]: E1203 00:24:31.571185 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.607318 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.607358 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.607367 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.607383 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.607393 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:31Z","lastTransitionTime":"2025-12-03T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.710727 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.710780 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.710791 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.710806 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.710820 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:31Z","lastTransitionTime":"2025-12-03T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.812652 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.812694 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.812703 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.812717 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.812727 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:31Z","lastTransitionTime":"2025-12-03T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.914564 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.914624 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.914635 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.914652 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:31 crc kubenswrapper[4912]: I1203 00:24:31.914665 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:31Z","lastTransitionTime":"2025-12-03T00:24:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.017175 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.017231 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.017239 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.017251 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.017261 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.118942 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.118971 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.118980 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.118992 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.119000 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.221469 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.221499 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.221506 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.221522 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.221535 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.323735 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.323819 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.323832 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.323848 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.323860 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.425385 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.425451 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.425463 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.425479 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.425491 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.528677 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.528721 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.528735 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.528750 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.528758 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.571161 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.571203 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.571247 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:32 crc kubenswrapper[4912]: E1203 00:24:32.571339 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:32 crc kubenswrapper[4912]: E1203 00:24:32.571484 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:32 crc kubenswrapper[4912]: E1203 00:24:32.571621 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.630662 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.630738 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.630762 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.630791 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.630809 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.733737 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.733784 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.733837 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.733852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.733864 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.836570 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.836604 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.836613 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.836627 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.836636 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.940270 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.940327 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.940344 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.940370 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:32 crc kubenswrapper[4912]: I1203 00:24:32.940399 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:32Z","lastTransitionTime":"2025-12-03T00:24:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.043204 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.043282 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.043300 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.043324 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.043343 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.146768 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.146846 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.146878 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.146910 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.146931 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.251661 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.251707 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.251719 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.251738 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.251753 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.354371 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.354409 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.354418 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.354468 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.354480 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.457842 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.457914 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.457933 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.457963 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.457983 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.561412 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.561506 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.561524 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.561551 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.561572 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.571105 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:33 crc kubenswrapper[4912]: E1203 00:24:33.571320 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.665364 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.665417 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.665448 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.665472 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.665485 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.768855 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.768913 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.768926 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.768941 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.768954 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.871197 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.871260 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.871274 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.871290 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.871305 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.974247 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.974315 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.974326 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.974340 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:33 crc kubenswrapper[4912]: I1203 00:24:33.974353 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:33Z","lastTransitionTime":"2025-12-03T00:24:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.077352 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.077406 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.077418 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.077464 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.077485 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:34Z","lastTransitionTime":"2025-12-03T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.180087 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.180128 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.180141 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.180157 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.180181 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:34Z","lastTransitionTime":"2025-12-03T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.282899 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.282941 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.282949 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.282965 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.282977 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:34Z","lastTransitionTime":"2025-12-03T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.385648 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.385687 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.385708 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.385720 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.385729 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:34Z","lastTransitionTime":"2025-12-03T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.488951 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.489020 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.489040 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.489069 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.489089 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:34Z","lastTransitionTime":"2025-12-03T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.571033 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.571112 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:34 crc kubenswrapper[4912]: E1203 00:24:34.571165 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.571467 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:34 crc kubenswrapper[4912]: E1203 00:24:34.571413 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:34 crc kubenswrapper[4912]: E1203 00:24:34.571522 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.588408 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.591309 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.591353 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.591367 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.591389 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.591403 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:34Z","lastTransitionTime":"2025-12-03T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.605684 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.614791 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.623906 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.636081 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.648501 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.661005 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.672353 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.690267 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e51ae1d1cae30a599494806c93191a2cf8c152c31bbb6af312262771a189e02\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"message\\\":\\\"3 00:24:08.095540 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-m2wz8\\\\nI1203 00:24:08.095548 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095557 6343 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-m2wz8 in node crc\\\\nI1203 00:24:08.095561 6343 obj_retry.go:365] Adding new object: *v1.Pod openshift-etcd/etcd-crc\\\\nI1203 00:24:08.095565 6343 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1203 00:24:08.095565 6343 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:08Z is\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.694370 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.694571 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.694705 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.694834 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.694954 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:34Z","lastTransitionTime":"2025-12-03T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.705575 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.726051 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.742278 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.757569 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.769616 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.780476 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.798113 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.798189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.798204 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.798225 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.798237 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:34Z","lastTransitionTime":"2025-12-03T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.803893 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.817980 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.831575 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:34Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.900266 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.900298 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.900307 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.900319 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:34 crc kubenswrapper[4912]: I1203 00:24:34.900330 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:34Z","lastTransitionTime":"2025-12-03T00:24:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.002233 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.002277 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.002287 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.002302 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.002314 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.105113 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.105185 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.105202 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.105235 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.105253 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.208131 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.208198 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.208213 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.208236 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.208250 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.311534 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.311583 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.311592 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.311609 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.311620 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.414194 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.414238 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.414249 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.414264 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.414273 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.517146 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.517187 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.517225 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.517242 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.517254 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.571052 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:35 crc kubenswrapper[4912]: E1203 00:24:35.571170 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.620235 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.620289 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.620301 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.620317 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.620327 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.723210 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.723249 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.723258 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.723273 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.723283 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.826094 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.826168 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.826202 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.826227 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.826240 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.929075 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.929117 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.929128 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.929145 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:35 crc kubenswrapper[4912]: I1203 00:24:35.929156 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:35Z","lastTransitionTime":"2025-12-03T00:24:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.031197 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.031297 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.031315 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.031338 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.031358 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.134167 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.134489 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.134499 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.134514 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.134526 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.237217 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.237259 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.237270 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.237286 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.237297 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.339509 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.339601 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.339609 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.339625 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.339635 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.442003 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.442051 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.442065 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.442083 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.442092 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.543973 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.544029 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.544041 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.544060 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.544073 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.570562 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.570660 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.570570 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:36 crc kubenswrapper[4912]: E1203 00:24:36.570717 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:36 crc kubenswrapper[4912]: E1203 00:24:36.570775 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:36 crc kubenswrapper[4912]: E1203 00:24:36.570863 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.647261 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.647305 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.647321 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.647344 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.647357 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.678327 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.678362 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.678371 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.678385 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.678394 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: E1203 00:24:36.689775 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:36Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.692983 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.693021 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.693033 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.693051 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.693067 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: E1203 00:24:36.710913 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:36Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.714726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.714759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.714770 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.714786 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.714798 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: E1203 00:24:36.729221 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:36Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.733062 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.733109 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.733122 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.733137 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.733148 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: E1203 00:24:36.749211 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:36Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.753153 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.753228 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.753259 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.753272 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.753282 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: E1203 00:24:36.765152 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:36Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:36 crc kubenswrapper[4912]: E1203 00:24:36.765305 4912 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.766779 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.766804 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.766815 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.766829 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.766841 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.869414 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.869508 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.869520 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.869553 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.869562 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.971990 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.972017 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.972027 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.972039 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:36 crc kubenswrapper[4912]: I1203 00:24:36.972048 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:36Z","lastTransitionTime":"2025-12-03T00:24:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.074525 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.074580 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.074592 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.074611 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.074624 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.176721 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.176776 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.176785 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.176799 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.176837 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.279701 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.279741 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.279749 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.279762 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.279771 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.382035 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.382106 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.382118 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.382135 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.382147 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.484549 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.484598 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.484611 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.484630 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.484643 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.571007 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:37 crc kubenswrapper[4912]: E1203 00:24:37.571128 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.572030 4912 scope.go:117] "RemoveContainer" containerID="e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63" Dec 03 00:24:37 crc kubenswrapper[4912]: E1203 00:24:37.572369 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.585321 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.586362 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.586403 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.586415 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.586451 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.586465 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.598467 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.612538 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.625769 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.642101 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.653423 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.671015 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.684066 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.688227 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.688262 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.688272 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.688287 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.688296 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.696901 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.707770 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.720173 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.731122 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.742912 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.763020 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.774709 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.785232 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.791292 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.791320 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.791331 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.791363 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.791374 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.797462 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.809807 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:37Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.894128 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.894166 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.894175 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.894190 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.894200 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.996327 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.996384 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.996399 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.996476 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:37 crc kubenswrapper[4912]: I1203 00:24:37.996490 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:37Z","lastTransitionTime":"2025-12-03T00:24:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.098804 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.098838 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.098851 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.098868 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.098881 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:38Z","lastTransitionTime":"2025-12-03T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.200752 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.200790 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.200800 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.200815 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.200827 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:38Z","lastTransitionTime":"2025-12-03T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.303184 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.303234 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.303244 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.303262 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.303273 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:38Z","lastTransitionTime":"2025-12-03T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.405601 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.405640 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.405648 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.405678 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.405688 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:38Z","lastTransitionTime":"2025-12-03T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.507533 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.507571 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.507582 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.507597 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.507607 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:38Z","lastTransitionTime":"2025-12-03T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.571485 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.571544 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.571613 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:38 crc kubenswrapper[4912]: E1203 00:24:38.571733 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:38 crc kubenswrapper[4912]: E1203 00:24:38.571877 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:38 crc kubenswrapper[4912]: E1203 00:24:38.571968 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.610015 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.610056 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.610064 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.610079 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.610089 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:38Z","lastTransitionTime":"2025-12-03T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.712850 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.712890 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.712900 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.712916 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.712929 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:38Z","lastTransitionTime":"2025-12-03T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.815762 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.815815 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.815829 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.815846 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.815858 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:38Z","lastTransitionTime":"2025-12-03T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.917749 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.917791 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.917799 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.917813 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:38 crc kubenswrapper[4912]: I1203 00:24:38.917822 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:38Z","lastTransitionTime":"2025-12-03T00:24:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.019776 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.019808 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.019818 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.019830 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.019839 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.121942 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.121989 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.122000 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.122013 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.122025 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.224819 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.224868 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.224876 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.224890 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.224900 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.326935 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.326997 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.327024 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.327041 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.327054 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.429523 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.429581 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.429594 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.429609 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.429619 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.531467 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.531513 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.531527 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.531543 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.531555 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.571140 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:39 crc kubenswrapper[4912]: E1203 00:24:39.571287 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.633964 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.634001 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.634010 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.634024 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.634034 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.736166 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.736216 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.736227 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.736245 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.736261 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.838619 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.838691 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.838703 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.838724 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.838735 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.940728 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.940770 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.940777 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.940791 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:39 crc kubenswrapper[4912]: I1203 00:24:39.940799 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:39Z","lastTransitionTime":"2025-12-03T00:24:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.042726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.042759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.042766 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.042778 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.042786 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.144836 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.144869 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.144879 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.144891 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.144899 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.247382 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.247424 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.247457 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.247477 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.247489 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.349874 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.349915 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.349924 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.349937 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.349945 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.451823 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.451863 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.451873 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.451887 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.451897 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.554817 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.554851 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.554860 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.554874 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.554883 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.571307 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.571365 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:40 crc kubenswrapper[4912]: E1203 00:24:40.571411 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.571307 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:40 crc kubenswrapper[4912]: E1203 00:24:40.571524 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:40 crc kubenswrapper[4912]: E1203 00:24:40.571595 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.657538 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.657575 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.657586 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.657794 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.657806 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.760225 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.760288 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.760299 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.760317 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.760329 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.862766 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.862822 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.862835 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.862852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.862865 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.966027 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.966066 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.966076 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.966112 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:40 crc kubenswrapper[4912]: I1203 00:24:40.966124 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:40Z","lastTransitionTime":"2025-12-03T00:24:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.068882 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.068921 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.068931 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.068946 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.068955 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.171651 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.171691 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.171700 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.171714 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.171724 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.274393 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.274452 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.274461 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.274474 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.274483 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.376823 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.376856 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.376864 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.376876 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.376886 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.384407 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:41 crc kubenswrapper[4912]: E1203 00:24:41.384589 4912 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:41 crc kubenswrapper[4912]: E1203 00:24:41.384656 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs podName:e49d5e0a-f59c-4dc1-9140-4b255576cd72 nodeName:}" failed. No retries permitted until 2025-12-03 00:25:13.38463907 +0000 UTC m=+99.026659630 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs") pod "network-metrics-daemon-kdznz" (UID: "e49d5e0a-f59c-4dc1-9140-4b255576cd72") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.479035 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.479074 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.479083 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.479095 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.479103 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.571418 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:41 crc kubenswrapper[4912]: E1203 00:24:41.571600 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.581281 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.581310 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.581321 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.581336 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.581347 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.683690 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.683726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.683734 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.683749 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.683758 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.785690 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.785726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.785735 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.785748 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.785758 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.888158 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.888208 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.888223 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.888239 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.888250 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.990894 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.990937 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.990946 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.990960 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:41 crc kubenswrapper[4912]: I1203 00:24:41.990969 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:41Z","lastTransitionTime":"2025-12-03T00:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.093065 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.093110 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.093119 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.093133 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.093143 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:42Z","lastTransitionTime":"2025-12-03T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.195757 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.195824 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.195843 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.195867 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.195884 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:42Z","lastTransitionTime":"2025-12-03T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.297475 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.297510 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.297521 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.297536 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.297546 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:42Z","lastTransitionTime":"2025-12-03T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.399611 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.399648 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.399658 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.399672 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.399681 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:42Z","lastTransitionTime":"2025-12-03T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.501829 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.501862 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.501873 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.501916 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.501928 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:42Z","lastTransitionTime":"2025-12-03T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.571292 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.571397 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.571295 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:42 crc kubenswrapper[4912]: E1203 00:24:42.571611 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:42 crc kubenswrapper[4912]: E1203 00:24:42.571653 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:42 crc kubenswrapper[4912]: E1203 00:24:42.571453 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.604741 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.604786 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.604797 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.604814 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.604829 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:42Z","lastTransitionTime":"2025-12-03T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.706740 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.706775 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.706785 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.706798 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.706806 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:42Z","lastTransitionTime":"2025-12-03T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.808849 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.808908 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.808920 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.808939 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.808952 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:42Z","lastTransitionTime":"2025-12-03T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.910686 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.910752 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.910776 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.910806 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:42 crc kubenswrapper[4912]: I1203 00:24:42.910828 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:42Z","lastTransitionTime":"2025-12-03T00:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.013996 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.014048 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.014065 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.014090 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.014107 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.116861 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.116902 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.116910 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.116924 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.116934 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.219844 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.219897 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.219910 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.219926 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.219937 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.322516 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.322580 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.322591 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.322631 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.322644 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.424777 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.424811 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.424820 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.424833 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.424843 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.526827 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.526867 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.526877 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.526892 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.526901 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.571544 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:43 crc kubenswrapper[4912]: E1203 00:24:43.571692 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.628973 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.629001 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.629012 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.629026 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.629035 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.731513 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.731552 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.731563 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.731578 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.731590 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.834452 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.834494 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.834504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.834517 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.834527 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.937078 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.937110 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.937119 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.937134 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.937147 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:43Z","lastTransitionTime":"2025-12-03T00:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.946510 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2wz8_8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77/kube-multus/0.log" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.946573 4912 generic.go:334] "Generic (PLEG): container finished" podID="8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77" containerID="d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6" exitCode=1 Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.946599 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2wz8" event={"ID":"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77","Type":"ContainerDied","Data":"d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6"} Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.946957 4912 scope.go:117] "RemoveContainer" containerID="d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.962719 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.976224 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:43 crc kubenswrapper[4912]: I1203 00:24:43.987808 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.001162 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:43Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.015099 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.027077 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.039647 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.039683 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.039694 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.039709 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.039719 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.045161 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.064051 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.075290 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.084216 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.092127 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.100160 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.110167 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.122846 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.131324 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.141330 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.141547 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.141635 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.141708 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.141766 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.143232 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.165510 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"2025-12-03T00:23:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428\\\\n2025-12-03T00:23:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428 to /host/opt/cni/bin/\\\\n2025-12-03T00:23:58Z [verbose] multus-daemon started\\\\n2025-12-03T00:23:58Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:24:43Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.174148 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.244211 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.244242 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.244253 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.244266 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.244276 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.346284 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.346311 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.346318 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.346329 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.346338 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.448539 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.448570 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.448585 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.448601 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.448612 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.550733 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.550763 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.550771 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.550786 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.550796 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.570769 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:44 crc kubenswrapper[4912]: E1203 00:24:44.570894 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.571049 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:44 crc kubenswrapper[4912]: E1203 00:24:44.571093 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.571345 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:44 crc kubenswrapper[4912]: E1203 00:24:44.571556 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.582221 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.595670 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.606217 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.619058 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.630349 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"2025-12-03T00:23:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428\\\\n2025-12-03T00:23:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428 to /host/opt/cni/bin/\\\\n2025-12-03T00:23:58Z [verbose] multus-daemon started\\\\n2025-12-03T00:23:58Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:24:43Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.642056 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.651945 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.651986 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.651997 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.652012 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.652024 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.659291 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.678844 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.689788 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.699854 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.710796 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.721020 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.730305 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.746852 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.754318 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.754353 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.754365 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.754381 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.754392 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.758328 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.769374 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.778641 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.789154 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.856666 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.856725 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.856742 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.856764 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.856780 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.951210 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2wz8_8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77/kube-multus/0.log" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.951261 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2wz8" event={"ID":"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77","Type":"ContainerStarted","Data":"62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.958748 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.958851 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.958864 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.958877 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.958887 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:44Z","lastTransitionTime":"2025-12-03T00:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.964231 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.974529 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:44 crc kubenswrapper[4912]: I1203 00:24:44.986739 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:44Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.005208 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.016774 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"2025-12-03T00:23:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428\\\\n2025-12-03T00:23:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428 to /host/opt/cni/bin/\\\\n2025-12-03T00:23:58Z [verbose] multus-daemon started\\\\n2025-12-03T00:23:58Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:24:43Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.028019 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.039950 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.051207 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.060510 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.060543 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.060553 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.060569 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.060577 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.062607 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.072648 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.091134 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.105960 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.118457 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.129298 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.139330 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.150089 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.160265 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.163614 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.163655 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.163665 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.163678 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.163689 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.178097 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:45Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.266337 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.266379 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.266390 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.266405 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.266416 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.369143 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.369185 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.369195 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.369211 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.369221 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.471853 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.471888 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.471896 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.471909 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.471918 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.571120 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:45 crc kubenswrapper[4912]: E1203 00:24:45.571265 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.573453 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.573479 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.573489 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.573502 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.573512 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.675479 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.675519 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.675529 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.675545 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.675558 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.777644 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.777678 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.777686 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.777699 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.777708 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.879981 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.880019 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.880029 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.880044 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.880055 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.982803 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.982853 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.982863 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.982881 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:45 crc kubenswrapper[4912]: I1203 00:24:45.982893 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:45Z","lastTransitionTime":"2025-12-03T00:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.084614 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.084862 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.085070 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.085243 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.085505 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.188053 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.188338 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.188420 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.188550 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.188619 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.290394 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.290422 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.290450 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.290468 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.290479 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.392750 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.392806 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.392815 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.392827 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.392836 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.494799 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.494848 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.494861 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.494877 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.494890 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.570544 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.570604 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.570544 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:46 crc kubenswrapper[4912]: E1203 00:24:46.570704 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:46 crc kubenswrapper[4912]: E1203 00:24:46.570730 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:46 crc kubenswrapper[4912]: E1203 00:24:46.570801 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.596755 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.596790 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.596799 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.596814 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.596823 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.699357 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.699420 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.699473 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.699492 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.699503 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.803273 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.803385 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.803406 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.803469 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.803500 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.905873 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.905909 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.905916 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.905929 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.905939 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.985818 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.985858 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.985869 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.985884 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:46 crc kubenswrapper[4912]: I1203 00:24:46.985894 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:46Z","lastTransitionTime":"2025-12-03T00:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:46 crc kubenswrapper[4912]: E1203 00:24:46.998643 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:46Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.002875 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.002988 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.003057 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.003160 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.003245 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: E1203 00:24:47.017653 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:47Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.021057 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.021089 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.021098 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.021111 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.021120 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: E1203 00:24:47.032692 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:47Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.035880 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.035913 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.035922 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.035937 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.035948 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: E1203 00:24:47.048139 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:47Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.051852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.051883 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.051892 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.051909 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.051918 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: E1203 00:24:47.063914 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:47Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:47 crc kubenswrapper[4912]: E1203 00:24:47.064026 4912 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.065408 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.065456 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.065469 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.065485 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.065496 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.167852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.167880 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.167888 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.167901 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.167910 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.270172 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.270201 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.270211 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.270223 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.270233 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.372512 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.372565 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.372575 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.372588 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.372597 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.474886 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.474946 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.474958 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.474974 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.474983 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.571021 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:47 crc kubenswrapper[4912]: E1203 00:24:47.571200 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.576946 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.577004 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.577014 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.577029 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.577041 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.679451 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.679520 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.679573 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.679590 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.679602 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.782457 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.782498 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.782510 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.782525 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.782534 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.884482 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.884512 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.884523 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.884539 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.884549 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.986847 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.986918 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.986929 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.986952 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:47 crc kubenswrapper[4912]: I1203 00:24:47.986964 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:47Z","lastTransitionTime":"2025-12-03T00:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.089325 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.089369 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.089378 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.089392 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.089401 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:48Z","lastTransitionTime":"2025-12-03T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.192322 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.192364 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.192378 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.192446 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.192461 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:48Z","lastTransitionTime":"2025-12-03T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.294553 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.294598 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.294609 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.294626 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.294637 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:48Z","lastTransitionTime":"2025-12-03T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.397756 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.397824 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.397836 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.397851 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.397862 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:48Z","lastTransitionTime":"2025-12-03T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.500188 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.500246 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.500255 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.500271 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.500281 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:48Z","lastTransitionTime":"2025-12-03T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.571491 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:48 crc kubenswrapper[4912]: E1203 00:24:48.571704 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.571722 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.571750 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:48 crc kubenswrapper[4912]: E1203 00:24:48.571999 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:48 crc kubenswrapper[4912]: E1203 00:24:48.572101 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.583777 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.605615 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.605660 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.605672 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.605686 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.605699 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:48Z","lastTransitionTime":"2025-12-03T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.707808 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.707855 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.707863 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.707877 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.707887 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:48Z","lastTransitionTime":"2025-12-03T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.810208 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.810260 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.810271 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.810287 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.810297 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:48Z","lastTransitionTime":"2025-12-03T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.912138 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.912183 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.912194 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.912211 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:48 crc kubenswrapper[4912]: I1203 00:24:48.912223 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:48Z","lastTransitionTime":"2025-12-03T00:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.015200 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.015265 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.015281 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.015307 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.015325 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.117824 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.117870 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.117880 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.117914 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.117925 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.220664 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.220708 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.220726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.220744 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.220759 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.323536 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.323608 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.323622 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.323658 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.323672 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.426823 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.426876 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.426889 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.426911 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.426926 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.530217 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.530263 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.530272 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.530287 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.530299 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.570918 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:49 crc kubenswrapper[4912]: E1203 00:24:49.571131 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.633192 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.633241 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.633253 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.633273 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.633286 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.735409 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.735500 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.735512 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.735527 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.735536 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.837483 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.837520 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.837531 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.837546 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.837557 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.940617 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.940691 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.940721 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.940756 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:49 crc kubenswrapper[4912]: I1203 00:24:49.940781 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:49Z","lastTransitionTime":"2025-12-03T00:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.043123 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.043173 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.043187 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.043206 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.043221 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.145031 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.145065 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.145073 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.145086 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.145095 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.247714 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.247837 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.247847 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.247862 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.247875 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.349858 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.349890 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.349898 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.349910 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.349918 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.452008 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.452063 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.452075 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.452089 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.452100 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.554276 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.554314 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.554325 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.554340 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.554353 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.570767 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.570797 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.570835 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:50 crc kubenswrapper[4912]: E1203 00:24:50.570889 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:50 crc kubenswrapper[4912]: E1203 00:24:50.570956 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:50 crc kubenswrapper[4912]: E1203 00:24:50.571013 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.657730 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.657758 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.657766 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.657801 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.657813 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.760282 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.760327 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.760338 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.760352 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.760362 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.863603 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.863653 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.863663 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.863676 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.863684 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.965770 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.965801 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.965809 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.965845 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:50 crc kubenswrapper[4912]: I1203 00:24:50.965857 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:50Z","lastTransitionTime":"2025-12-03T00:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.068540 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.068591 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.068603 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.068618 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.068632 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.171385 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.171484 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.171495 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.171509 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.171518 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.274139 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.274185 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.274196 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.274211 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.274220 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.379545 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.379594 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.379603 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.379625 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.379636 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.482610 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.482655 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.482667 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.482685 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.482697 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.570526 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:51 crc kubenswrapper[4912]: E1203 00:24:51.570670 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.571382 4912 scope.go:117] "RemoveContainer" containerID="e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.585100 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.585143 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.585156 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.585172 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.585183 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.687619 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.687669 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.687684 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.687705 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.687722 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.790656 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.790708 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.790720 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.790739 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.790751 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.893896 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.893939 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.893949 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.893966 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.893976 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.997444 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.997509 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.997525 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.997581 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:51 crc kubenswrapper[4912]: I1203 00:24:51.997598 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:51Z","lastTransitionTime":"2025-12-03T00:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.101716 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.101753 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.101763 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.101783 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.101793 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:52Z","lastTransitionTime":"2025-12-03T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.205394 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.205449 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.205462 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.205479 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.205491 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:52Z","lastTransitionTime":"2025-12-03T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.308147 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.308184 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.308195 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.308213 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.308225 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:52Z","lastTransitionTime":"2025-12-03T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.410598 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.410632 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.410642 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.410654 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.410665 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:52Z","lastTransitionTime":"2025-12-03T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.512941 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.512986 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.512998 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.513012 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.513023 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:52Z","lastTransitionTime":"2025-12-03T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.570644 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.570679 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.570764 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:52 crc kubenswrapper[4912]: E1203 00:24:52.570820 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:52 crc kubenswrapper[4912]: E1203 00:24:52.570921 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:52 crc kubenswrapper[4912]: E1203 00:24:52.571059 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.615675 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.615725 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.615737 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.615753 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.615766 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:52Z","lastTransitionTime":"2025-12-03T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.718175 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.718213 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.718223 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.718237 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.718248 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:52Z","lastTransitionTime":"2025-12-03T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.820573 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.820612 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.820623 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.820637 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.820646 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:52Z","lastTransitionTime":"2025-12-03T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.922529 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.922568 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.922576 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.922589 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.922598 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:52Z","lastTransitionTime":"2025-12-03T00:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.974100 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/2.log" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.976247 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.976681 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:24:52 crc kubenswrapper[4912]: I1203 00:24:52.994132 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:52Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.005992 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.018846 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.024363 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.024410 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.024422 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.024455 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.024466 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.028362 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.036820 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.046826 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.057689 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.067583 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.077706 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34bbddb7-1cd0-4afc-afca-fad3ceb28d61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://089195fc4e809a322be7b9f5a536d7fb8924fad3a5ad89ec6dd68994dd0e9646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.089061 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.103304 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"2025-12-03T00:23:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428\\\\n2025-12-03T00:23:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428 to /host/opt/cni/bin/\\\\n2025-12-03T00:23:58Z [verbose] multus-daemon started\\\\n2025-12-03T00:23:58Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:24:43Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.114103 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.126459 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.126506 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.126515 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.126529 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.126539 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.127447 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.138544 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.151987 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.163037 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.174270 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.195008 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.209333 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:53Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.229083 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.229121 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.229130 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.229144 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.229154 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.330973 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.331022 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.331032 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.331046 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.331056 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.433387 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.433420 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.433442 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.433459 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.433470 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.535658 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.535726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.535739 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.535756 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.535812 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.570967 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:53 crc kubenswrapper[4912]: E1203 00:24:53.571114 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.638747 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.638795 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.638806 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.638823 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.638834 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.742007 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.742058 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.742069 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.742096 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.742110 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.844692 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.844728 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.844737 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.844752 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.844762 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.947058 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.947104 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.947112 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.947128 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.947147 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:53Z","lastTransitionTime":"2025-12-03T00:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.982321 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/3.log" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.983414 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/2.log" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.986540 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" exitCode=1 Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.986605 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.986706 4912 scope.go:117] "RemoveContainer" containerID="e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63" Dec 03 00:24:53 crc kubenswrapper[4912]: I1203 00:24:53.987578 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:24:53 crc kubenswrapper[4912]: E1203 00:24:53.987796 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.013663 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.029214 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"2025-12-03T00:23:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428\\\\n2025-12-03T00:23:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428 to /host/opt/cni/bin/\\\\n2025-12-03T00:23:58Z [verbose] multus-daemon started\\\\n2025-12-03T00:23:58Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:24:43Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.040733 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.050248 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.050291 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.050327 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.050351 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.050366 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.053329 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34bbddb7-1cd0-4afc-afca-fad3ceb28d61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://089195fc4e809a322be7b9f5a536d7fb8924fad3a5ad89ec6dd68994dd0e9646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.067746 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.083655 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.098209 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.112131 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.131051 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:53Z\\\",\\\"message\\\":\\\"t: *v1.Pod openshift-multus/network-metrics-daemon-kdznz\\\\nI1203 00:24:53.061488 6904 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 in node crc\\\\nI1203 00:24:53.061567 6904 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-kdznz in node crc\\\\nI1203 00:24:53.061573 6904 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 after 0 failed attempt(s)\\\\nF1203 00:24:53.061578 6904 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:52Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:53.061\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.144785 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.153470 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.153527 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.153539 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.153559 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.153572 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.157501 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.170292 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.182760 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.195268 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.204768 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.223357 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.237109 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.246576 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.256080 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.256110 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.256119 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.256133 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.256143 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.257025 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.358940 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.358990 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.359002 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.359022 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.359034 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.461660 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.461697 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.461708 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.461724 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.461733 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.563677 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.563726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.563741 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.563757 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.563770 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.571023 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.571124 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:54 crc kubenswrapper[4912]: E1203 00:24:54.571239 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.571498 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:54 crc kubenswrapper[4912]: E1203 00:24:54.571578 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:54 crc kubenswrapper[4912]: E1203 00:24:54.571754 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.584530 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.599031 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.615090 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.635347 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2e010f679a8f5a128d63c836a6d55829a6ad82f7df4abe2035ab69933b75a63\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:24Z\\\",\\\"message\\\":\\\"od openshift-etcd/etcd-crc\\\\nI1203 00:24:24.548919 6547 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 00:24:24.548864 6547 lb_config.go:1031] Cluster endpoints for openshift-operator-lifecycle-manager/olm-operator-metrics for network=default are: map[]\\\\nF1203 00:24:24.548931 6547 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:24Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:24.548935 6547 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:53Z\\\",\\\"message\\\":\\\"t: *v1.Pod openshift-multus/network-metrics-daemon-kdznz\\\\nI1203 00:24:53.061488 6904 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 in node crc\\\\nI1203 00:24:53.061567 6904 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-kdznz in node crc\\\\nI1203 00:24:53.061573 6904 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 after 0 failed attempt(s)\\\\nF1203 00:24:53.061578 6904 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:52Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:53.061\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.657039 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.666749 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.666810 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.666824 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.666847 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.666862 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.671544 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.685725 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.697398 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.710785 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.722717 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.742089 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.756007 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.767235 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.769628 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.769668 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.769678 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.769717 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.769728 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.779634 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.791629 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.804508 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"2025-12-03T00:23:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428\\\\n2025-12-03T00:23:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428 to /host/opt/cni/bin/\\\\n2025-12-03T00:23:58Z [verbose] multus-daemon started\\\\n2025-12-03T00:23:58Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:24:43Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.816986 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.827058 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34bbddb7-1cd0-4afc-afca-fad3ceb28d61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://089195fc4e809a322be7b9f5a536d7fb8924fad3a5ad89ec6dd68994dd0e9646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.843968 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:54Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.872266 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.872315 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.872325 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.872341 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.872353 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.975589 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.975657 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.975671 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.975695 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.975709 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:54Z","lastTransitionTime":"2025-12-03T00:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.993707 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/3.log" Dec 03 00:24:54 crc kubenswrapper[4912]: I1203 00:24:54.998830 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:24:54 crc kubenswrapper[4912]: E1203 00:24:54.999147 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.016385 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.035596 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.048840 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.063007 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34bbddb7-1cd0-4afc-afca-fad3ceb28d61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://089195fc4e809a322be7b9f5a536d7fb8924fad3a5ad89ec6dd68994dd0e9646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.079029 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.079069 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.079080 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.079097 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.079109 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:55Z","lastTransitionTime":"2025-12-03T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.079473 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.096910 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"2025-12-03T00:23:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428\\\\n2025-12-03T00:23:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428 to /host/opt/cni/bin/\\\\n2025-12-03T00:23:58Z [verbose] multus-daemon started\\\\n2025-12-03T00:23:58Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:24:43Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.113582 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.131253 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.149291 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.166418 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.183049 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.183100 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.183109 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.183127 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.183141 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:55Z","lastTransitionTime":"2025-12-03T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.193890 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.215397 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.236091 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:53Z\\\",\\\"message\\\":\\\"t: *v1.Pod openshift-multus/network-metrics-daemon-kdznz\\\\nI1203 00:24:53.061488 6904 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 in node crc\\\\nI1203 00:24:53.061567 6904 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-kdznz in node crc\\\\nI1203 00:24:53.061573 6904 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 after 0 failed attempt(s)\\\\nF1203 00:24:53.061578 6904 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:52Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:53.061\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.254201 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.279608 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.286409 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.286504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.286518 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.286542 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.286560 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:55Z","lastTransitionTime":"2025-12-03T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.295980 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.309248 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.322932 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.338401 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:55Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.390469 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.390517 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.390530 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.390549 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.390562 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:55Z","lastTransitionTime":"2025-12-03T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.493705 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.493744 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.493755 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.493770 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.493782 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:55Z","lastTransitionTime":"2025-12-03T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.571675 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:55 crc kubenswrapper[4912]: E1203 00:24:55.571920 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.597481 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.597561 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.597573 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.597596 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.597609 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:55Z","lastTransitionTime":"2025-12-03T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.701227 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.701302 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.701317 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.701346 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.701362 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:55Z","lastTransitionTime":"2025-12-03T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.805788 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.805839 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.805852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.805871 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.805885 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:55Z","lastTransitionTime":"2025-12-03T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.909052 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.909108 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.909119 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.909136 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:55 crc kubenswrapper[4912]: I1203 00:24:55.909147 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:55Z","lastTransitionTime":"2025-12-03T00:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.012065 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.012189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.012212 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.012243 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.012266 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.114360 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.114397 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.114406 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.114460 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.114476 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.216839 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.216906 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.216924 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.216945 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.216963 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.319872 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.319918 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.319928 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.319959 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.319972 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.423079 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.423129 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.423140 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.423155 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.423167 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.432504 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.432604 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.432651 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.432769 4912 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.432813 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:26:00.43280091 +0000 UTC m=+146.074821470 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.432860 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:00.432854021 +0000 UTC m=+146.074874581 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.432939 4912 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.432981 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 00:26:00.432974374 +0000 UTC m=+146.074994934 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.526961 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.527230 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.527298 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.527365 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.527463 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.533918 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.534066 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.534253 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.534342 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.534422 4912 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.534538 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 00:26:00.53451886 +0000 UTC m=+146.176539420 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.534291 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.534785 4912 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.534844 4912 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.534930 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 00:26:00.534918992 +0000 UTC m=+146.176939542 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.571301 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.571471 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.571326 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.571687 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.571935 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:56 crc kubenswrapper[4912]: E1203 00:24:56.571841 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.630829 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.630865 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.630875 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.630888 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.630897 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.734379 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.734707 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.734775 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.734841 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.734894 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.837244 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.837310 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.837323 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.837343 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.837356 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.940299 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.940340 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.940350 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.940365 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:56 crc kubenswrapper[4912]: I1203 00:24:56.940378 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:56Z","lastTransitionTime":"2025-12-03T00:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.042145 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.042203 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.042214 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.042233 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.042247 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.081022 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.081066 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.081075 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.081089 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.081097 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: E1203 00:24:57.092737 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.095901 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.095936 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.095946 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.095960 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.095970 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: E1203 00:24:57.111009 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.114524 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.114561 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.114571 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.114585 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.114596 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: E1203 00:24:57.126677 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.129649 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.129690 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.129707 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.129729 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.129741 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: E1203 00:24:57.143044 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.146420 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.146533 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.146612 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.146686 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.146749 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: E1203 00:24:57.157744 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:57Z is after 2025-08-24T17:21:41Z" Dec 03 00:24:57 crc kubenswrapper[4912]: E1203 00:24:57.157853 4912 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.159088 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.159130 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.159142 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.159159 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.159172 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.261175 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.261230 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.261246 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.261268 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.261286 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.364150 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.364194 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.364205 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.364223 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.364235 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.469079 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.469152 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.469169 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.469190 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.469210 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.570491 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:57 crc kubenswrapper[4912]: E1203 00:24:57.570624 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.572094 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.572157 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.572175 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.572190 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.572203 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.674872 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.674936 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.674948 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.674978 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.674989 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.777463 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.777496 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.777505 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.777536 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.777546 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.880039 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.880071 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.880080 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.880094 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.880102 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.981827 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.981863 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.981871 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.981884 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:57 crc kubenswrapper[4912]: I1203 00:24:57.981893 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:57Z","lastTransitionTime":"2025-12-03T00:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.084405 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.084483 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.084501 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.084523 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.084539 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:58Z","lastTransitionTime":"2025-12-03T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.186280 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.186325 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.186336 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.186352 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.186366 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:58Z","lastTransitionTime":"2025-12-03T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.288574 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.288638 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.288662 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.288693 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.288713 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:58Z","lastTransitionTime":"2025-12-03T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.390626 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.390666 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.390678 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.390696 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.390708 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:58Z","lastTransitionTime":"2025-12-03T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.493177 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.493220 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.493233 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.493249 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.493260 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:58Z","lastTransitionTime":"2025-12-03T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.571663 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.571683 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.572230 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:24:58 crc kubenswrapper[4912]: E1203 00:24:58.572384 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:24:58 crc kubenswrapper[4912]: E1203 00:24:58.572508 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:24:58 crc kubenswrapper[4912]: E1203 00:24:58.572509 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.595027 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.595069 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.595086 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.595106 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.595120 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:58Z","lastTransitionTime":"2025-12-03T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.697574 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.697615 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.697648 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.697662 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.697671 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:58Z","lastTransitionTime":"2025-12-03T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.800023 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.800066 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.800078 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.800102 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.800127 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:58Z","lastTransitionTime":"2025-12-03T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.902798 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.902834 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.902844 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.902859 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:58 crc kubenswrapper[4912]: I1203 00:24:58.902870 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:58Z","lastTransitionTime":"2025-12-03T00:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.004887 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.004930 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.004945 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.004966 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.004983 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.107991 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.108364 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.108603 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.108978 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.109334 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.211711 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.211794 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.211828 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.211855 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.211876 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.314711 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.314774 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.314792 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.314816 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.314833 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.417518 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.417552 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.417564 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.417578 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.417588 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.519548 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.519582 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.519589 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.519602 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.519611 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.571231 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:24:59 crc kubenswrapper[4912]: E1203 00:24:59.571547 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.622050 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.622090 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.622101 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.622118 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.622129 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.725605 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.725658 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.725670 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.725688 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.725701 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.828864 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.829184 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.829197 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.829212 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.829224 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.931995 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.932046 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.932058 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.932071 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:24:59 crc kubenswrapper[4912]: I1203 00:24:59.932080 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:24:59Z","lastTransitionTime":"2025-12-03T00:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.035339 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.035386 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.035397 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.035413 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.035444 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.137497 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.137534 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.137544 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.137562 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.137575 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.239756 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.239785 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.239793 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.239804 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.239812 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.342122 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.342156 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.342164 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.342176 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.342185 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.444773 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.444822 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.444832 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.444846 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.444855 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.547504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.547545 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.547553 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.547567 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.547576 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.571203 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.571201 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:00 crc kubenswrapper[4912]: E1203 00:25:00.571349 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.571222 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:00 crc kubenswrapper[4912]: E1203 00:25:00.571421 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:00 crc kubenswrapper[4912]: E1203 00:25:00.571521 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.649323 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.649386 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.649404 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.649491 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.649508 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.752265 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.752310 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.752321 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.752336 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.752349 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.854490 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.854535 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.854568 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.854583 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.854593 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.957086 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.957164 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.957188 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.957221 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:00 crc kubenswrapper[4912]: I1203 00:25:00.957243 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:00Z","lastTransitionTime":"2025-12-03T00:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.060164 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.060249 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.060267 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.060326 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.060346 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.163106 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.163137 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.163146 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.163159 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.163168 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.266390 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.266477 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.266494 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.266514 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.266527 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.369618 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.369676 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.369691 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.369716 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.369735 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.472703 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.472752 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.472761 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.472777 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.472787 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.570843 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:01 crc kubenswrapper[4912]: E1203 00:25:01.571057 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.575769 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.575815 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.575832 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.575856 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.575872 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.678555 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.678594 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.678603 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.678616 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.678625 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.781765 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.781839 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.781852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.781879 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.781902 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.884831 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.884914 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.884930 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.884952 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.884969 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.987793 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.987830 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.987838 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.987852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:01 crc kubenswrapper[4912]: I1203 00:25:01.987861 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:01Z","lastTransitionTime":"2025-12-03T00:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.090080 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.090115 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.090124 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.090138 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.090148 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:02Z","lastTransitionTime":"2025-12-03T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.191892 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.191984 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.191998 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.192015 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.192026 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:02Z","lastTransitionTime":"2025-12-03T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.294108 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.294151 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.294163 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.294177 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.294190 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:02Z","lastTransitionTime":"2025-12-03T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.398236 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.398279 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.398289 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.398303 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.398312 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:02Z","lastTransitionTime":"2025-12-03T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.499957 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.499988 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.499999 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.500015 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.500027 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:02Z","lastTransitionTime":"2025-12-03T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.571117 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.571186 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.571244 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:02 crc kubenswrapper[4912]: E1203 00:25:02.571346 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:02 crc kubenswrapper[4912]: E1203 00:25:02.571489 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:02 crc kubenswrapper[4912]: E1203 00:25:02.571552 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.602551 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.602584 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.602594 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.602608 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.602618 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:02Z","lastTransitionTime":"2025-12-03T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.705143 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.705170 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.705178 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.705190 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.705198 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:02Z","lastTransitionTime":"2025-12-03T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.807054 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.807112 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.807129 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.807152 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.807169 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:02Z","lastTransitionTime":"2025-12-03T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.908859 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.908904 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.908917 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.908930 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:02 crc kubenswrapper[4912]: I1203 00:25:02.908940 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:02Z","lastTransitionTime":"2025-12-03T00:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.011412 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.011466 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.011475 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.011488 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.011500 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.113377 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.113408 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.113480 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.113496 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.113505 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.215985 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.216024 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.216041 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.216064 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.216076 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.318223 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.318268 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.318282 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.318300 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.318313 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.420123 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.420166 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.420175 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.420188 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.420197 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.522244 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.522326 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.522346 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.522372 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.522393 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.571449 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:03 crc kubenswrapper[4912]: E1203 00:25:03.571575 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.625227 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.625270 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.625281 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.625298 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.625309 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.727626 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.727674 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.727690 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.727706 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.727719 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.829983 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.830021 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.830032 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.830048 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.830062 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.932178 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.932220 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.932228 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.932240 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:03 crc kubenswrapper[4912]: I1203 00:25:03.932250 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:03Z","lastTransitionTime":"2025-12-03T00:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.034672 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.034718 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.034730 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.034747 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.034760 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.136546 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.136581 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.136592 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.136633 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.136642 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.238814 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.238850 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.238859 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.238875 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.238883 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.341497 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.341535 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.341544 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.341557 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.341566 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.443673 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.443710 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.443717 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.443748 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.443759 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.545696 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.545739 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.545752 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.545769 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.545781 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.570968 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:04 crc kubenswrapper[4912]: E1203 00:25:04.571088 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.571156 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.571241 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:04 crc kubenswrapper[4912]: E1203 00:25:04.571260 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:04 crc kubenswrapper[4912]: E1203 00:25:04.571381 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.586555 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.605774 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:53Z\\\",\\\"message\\\":\\\"t: *v1.Pod openshift-multus/network-metrics-daemon-kdznz\\\\nI1203 00:24:53.061488 6904 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 in node crc\\\\nI1203 00:24:53.061567 6904 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-kdznz in node crc\\\\nI1203 00:24:53.061573 6904 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 after 0 failed attempt(s)\\\\nF1203 00:24:53.061578 6904 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:52Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:53.061\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.622765 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.636684 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.648965 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.648958 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.649025 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.649157 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.649180 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.649193 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.667534 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.680306 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.690159 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.707787 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.722125 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.733312 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.744961 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.751752 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.751802 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.751818 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.751838 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.751852 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.758546 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.772067 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.781924 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.789802 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34bbddb7-1cd0-4afc-afca-fad3ceb28d61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://089195fc4e809a322be7b9f5a536d7fb8924fad3a5ad89ec6dd68994dd0e9646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.800343 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.810715 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"2025-12-03T00:23:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428\\\\n2025-12-03T00:23:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428 to /host/opt/cni/bin/\\\\n2025-12-03T00:23:58Z [verbose] multus-daemon started\\\\n2025-12-03T00:23:58Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:24:43Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.820388 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:04Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.853762 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.853792 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.853800 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.853812 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.853820 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.955320 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.955362 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.955375 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.955390 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:04 crc kubenswrapper[4912]: I1203 00:25:04.955402 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:04Z","lastTransitionTime":"2025-12-03T00:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.057948 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.058185 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.058260 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.058347 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.058423 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.161255 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.161305 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.161317 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.161337 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.161348 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.263731 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.263772 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.263784 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.263798 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.263808 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.365559 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.365601 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.365612 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.365628 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.365639 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.467783 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.467823 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.467835 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.467851 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.467861 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.570253 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.570285 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.570294 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.570307 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.570317 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.570526 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:05 crc kubenswrapper[4912]: E1203 00:25:05.570633 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.672536 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.672584 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.672595 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.672610 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.672621 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.775117 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.775167 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.775181 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.775197 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.775211 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.877164 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.877198 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.877206 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.877219 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.877228 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.979895 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.980155 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.980220 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.980288 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:05 crc kubenswrapper[4912]: I1203 00:25:05.980350 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:05Z","lastTransitionTime":"2025-12-03T00:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.083222 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.083269 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.083278 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.083293 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.083305 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:06Z","lastTransitionTime":"2025-12-03T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.186246 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.186292 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.186304 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.186322 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.186399 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:06Z","lastTransitionTime":"2025-12-03T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.290278 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.290366 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.290380 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.290404 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.290420 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:06Z","lastTransitionTime":"2025-12-03T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.393828 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.393902 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.393917 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.393942 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.393958 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:06Z","lastTransitionTime":"2025-12-03T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.497636 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.497682 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.497692 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.497710 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.497720 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:06Z","lastTransitionTime":"2025-12-03T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.571234 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.571338 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:06 crc kubenswrapper[4912]: E1203 00:25:06.571547 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.571571 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:06 crc kubenswrapper[4912]: E1203 00:25:06.571675 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:06 crc kubenswrapper[4912]: E1203 00:25:06.571844 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.573076 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:25:06 crc kubenswrapper[4912]: E1203 00:25:06.573361 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.601588 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.601662 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.601681 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.601710 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.601732 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:06Z","lastTransitionTime":"2025-12-03T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.704173 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.704220 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.704234 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.704252 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.704267 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:06Z","lastTransitionTime":"2025-12-03T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.807145 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.807187 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.807198 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.807212 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.807223 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:06Z","lastTransitionTime":"2025-12-03T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.909482 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.909557 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.909574 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.909599 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:06 crc kubenswrapper[4912]: I1203 00:25:06.909617 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:06Z","lastTransitionTime":"2025-12-03T00:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.012235 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.012298 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.012311 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.012331 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.012347 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.115284 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.115325 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.115336 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.115359 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.115373 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.218946 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.219000 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.219011 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.219029 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.219046 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.322877 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.322965 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.322980 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.322997 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.323007 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.426021 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.426063 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.426075 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.426089 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.426099 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.485028 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.485094 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.485106 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.485124 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.485135 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: E1203 00:25:07.503945 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.509860 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.509919 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.509936 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.509980 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.510004 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: E1203 00:25:07.529799 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.535314 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.535414 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.535475 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.535500 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.535522 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: E1203 00:25:07.555779 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.561555 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.561622 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.561641 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.561667 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.561686 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.570943 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:07 crc kubenswrapper[4912]: E1203 00:25:07.571194 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:07 crc kubenswrapper[4912]: E1203 00:25:07.578146 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.584654 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.584711 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.584728 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.584750 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.584767 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: E1203 00:25:07.601395 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:07Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:07 crc kubenswrapper[4912]: E1203 00:25:07.601648 4912 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.604355 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.604410 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.604422 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.604478 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.604488 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.707310 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.707367 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.707379 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.707399 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.707413 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.810537 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.810571 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.810579 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.810592 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.810600 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.913552 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.913626 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.913683 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.913703 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:07 crc kubenswrapper[4912]: I1203 00:25:07.913713 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:07Z","lastTransitionTime":"2025-12-03T00:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.016131 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.016173 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.016185 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.016200 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.016212 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.118759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.118801 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.118809 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.118823 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.118835 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.220907 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.220960 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.220971 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.220988 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.221000 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.323096 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.323130 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.323139 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.323151 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.323162 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.426277 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.426320 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.426329 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.426346 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.426356 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.528882 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.528918 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.528926 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.528940 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.528949 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.570541 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.570584 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.570749 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:08 crc kubenswrapper[4912]: E1203 00:25:08.570892 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:08 crc kubenswrapper[4912]: E1203 00:25:08.571034 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:08 crc kubenswrapper[4912]: E1203 00:25:08.571145 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.630878 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.631172 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.631335 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.631406 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.631497 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.734077 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.734113 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.734125 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.734161 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.734177 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.837360 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.837405 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.837414 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.837445 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.837455 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.944311 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.944354 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.944365 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.944380 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:08 crc kubenswrapper[4912]: I1203 00:25:08.944389 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:08Z","lastTransitionTime":"2025-12-03T00:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.046557 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.046596 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.046605 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.046620 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.046633 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.149781 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.149850 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.149868 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.149897 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.149916 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.253124 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.253206 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.253229 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.253264 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.253287 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.356730 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.356796 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.356818 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.356848 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.356868 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.459148 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.459189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.459199 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.459213 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.459224 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.561257 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.561290 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.561298 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.561310 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.561318 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.570695 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:09 crc kubenswrapper[4912]: E1203 00:25:09.571006 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.663990 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.664027 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.664040 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.664057 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.664070 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.766314 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.766351 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.766360 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.766375 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.766384 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.868681 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.868754 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.868780 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.868805 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.868821 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.970950 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.971010 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.971024 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.971045 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:09 crc kubenswrapper[4912]: I1203 00:25:09.971060 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:09Z","lastTransitionTime":"2025-12-03T00:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.073679 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.073722 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.073740 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.073757 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.073766 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.176217 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.176247 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.176255 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.176268 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.176276 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.278139 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.278179 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.278197 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.278213 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.278224 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.380926 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.380962 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.380971 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.380985 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.380995 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.482919 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.482985 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.482997 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.483013 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.483025 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.570616 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.570659 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.570674 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:10 crc kubenswrapper[4912]: E1203 00:25:10.570740 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:10 crc kubenswrapper[4912]: E1203 00:25:10.570821 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:10 crc kubenswrapper[4912]: E1203 00:25:10.570938 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.585640 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.585681 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.585692 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.585709 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.585722 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.687658 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.687982 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.687994 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.688009 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.688019 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.790029 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.790079 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.790090 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.790108 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.790119 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.892554 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.892591 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.892607 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.892623 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.892642 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.995495 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.995544 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.995559 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.995582 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:10 crc kubenswrapper[4912]: I1203 00:25:10.995598 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:10Z","lastTransitionTime":"2025-12-03T00:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.099061 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.099106 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.099119 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.099135 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.099146 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:11Z","lastTransitionTime":"2025-12-03T00:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.201601 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.201670 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.201694 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.201717 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.201750 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:11Z","lastTransitionTime":"2025-12-03T00:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.303729 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.303787 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.303803 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.303824 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.303838 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:11Z","lastTransitionTime":"2025-12-03T00:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.406949 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.407000 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.407014 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.407028 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.407037 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:11Z","lastTransitionTime":"2025-12-03T00:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.509822 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.509876 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.509890 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.509909 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.509922 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:11Z","lastTransitionTime":"2025-12-03T00:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.570509 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:11 crc kubenswrapper[4912]: E1203 00:25:11.570635 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.612189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.612224 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.612233 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.612245 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.612254 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:11Z","lastTransitionTime":"2025-12-03T00:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.714631 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.714673 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.714683 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.714696 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.714705 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:11Z","lastTransitionTime":"2025-12-03T00:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.816617 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.816660 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.816672 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.816687 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.816697 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:11Z","lastTransitionTime":"2025-12-03T00:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.920058 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.920146 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.920170 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.920203 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:11 crc kubenswrapper[4912]: I1203 00:25:11.920228 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:11Z","lastTransitionTime":"2025-12-03T00:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.023800 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.023848 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.023857 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.023869 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.023881 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.126632 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.126679 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.126690 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.126704 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.126712 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.229671 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.229718 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.229733 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.229752 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.229768 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.333354 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.333395 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.333406 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.333452 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.333464 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.436173 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.436242 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.436264 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.436292 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.436313 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.539261 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.539319 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.539331 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.539345 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.539356 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.570828 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.570883 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:12 crc kubenswrapper[4912]: E1203 00:25:12.571053 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.571114 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:12 crc kubenswrapper[4912]: E1203 00:25:12.571249 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:12 crc kubenswrapper[4912]: E1203 00:25:12.571380 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.642369 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.642495 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.642519 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.642552 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.642573 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.745877 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.745918 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.745934 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.745954 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.745968 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.848582 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.848643 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.848680 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.848712 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.848787 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.951679 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.951706 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.951714 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.951726 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:12 crc kubenswrapper[4912]: I1203 00:25:12.951735 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:12Z","lastTransitionTime":"2025-12-03T00:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.054796 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.054852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.054867 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.054885 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.054904 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.158366 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.158408 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.158417 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.158446 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.158457 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.261625 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.261668 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.261678 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.261693 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.261703 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.364798 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.364841 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.364851 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.364869 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.364881 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.409122 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:13 crc kubenswrapper[4912]: E1203 00:25:13.409402 4912 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:25:13 crc kubenswrapper[4912]: E1203 00:25:13.409547 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs podName:e49d5e0a-f59c-4dc1-9140-4b255576cd72 nodeName:}" failed. No retries permitted until 2025-12-03 00:26:17.409514218 +0000 UTC m=+163.051534818 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs") pod "network-metrics-daemon-kdznz" (UID: "e49d5e0a-f59c-4dc1-9140-4b255576cd72") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.468153 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.468226 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.468250 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.468285 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.468305 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.571210 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:13 crc kubenswrapper[4912]: E1203 00:25:13.571365 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.571654 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.571739 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.571750 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.571774 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.571788 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.675215 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.675281 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.675305 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.675337 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.675360 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.778201 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.778267 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.778290 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.778321 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.778341 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.880887 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.880936 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.880944 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.880958 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.880967 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.983054 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.983093 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.983103 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.983117 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:13 crc kubenswrapper[4912]: I1203 00:25:13.983128 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:13Z","lastTransitionTime":"2025-12-03T00:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.085633 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.085674 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.085684 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.085697 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.085710 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:14Z","lastTransitionTime":"2025-12-03T00:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.187888 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.187928 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.187941 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.187960 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.187975 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:14Z","lastTransitionTime":"2025-12-03T00:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.290964 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.291003 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.291013 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.291027 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.291067 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:14Z","lastTransitionTime":"2025-12-03T00:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.393138 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.393167 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.393175 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.393188 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.393197 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:14Z","lastTransitionTime":"2025-12-03T00:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.495675 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.495742 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.495753 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.495769 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.495780 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:14Z","lastTransitionTime":"2025-12-03T00:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.570746 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.570825 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:14 crc kubenswrapper[4912]: E1203 00:25:14.570873 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:14 crc kubenswrapper[4912]: E1203 00:25:14.570971 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.571095 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:14 crc kubenswrapper[4912]: E1203 00:25:14.571300 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.590498 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34bbddb7-1cd0-4afc-afca-fad3ceb28d61\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://089195fc4e809a322be7b9f5a536d7fb8924fad3a5ad89ec6dd68994dd0e9646\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4a3181737f9219156382c0de580bf956ec0f23195bdd7f32975ce1b9be36e86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.598859 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.598908 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.598920 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.598935 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.598949 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:14Z","lastTransitionTime":"2025-12-03T00:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.603357 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bf502f4-28d8-4d3e-bbfc-ebd54758b252\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 00:23:46.831981 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 00:23:46.834320 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2535086094/tls.crt::/tmp/serving-cert-2535086094/tls.key\\\\\\\"\\\\nI1203 00:23:52.461321 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 00:23:52.463165 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 00:23:52.463214 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 00:23:52.463255 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 00:23:52.463288 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 00:23:52.467754 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 00:23:52.467810 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467816 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 00:23:52.467822 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 00:23:52.467826 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 00:23:52.467830 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 00:23:52.467833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 00:23:52.468073 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 00:23:52.469966 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.622705 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m2wz8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:43Z\\\",\\\"message\\\":\\\"2025-12-03T00:23:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428\\\\n2025-12-03T00:23:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_06c2d0c1-2fae-4dd7-ac34-eed00ca38428 to /host/opt/cni/bin/\\\\n2025-12-03T00:23:58Z [verbose] multus-daemon started\\\\n2025-12-03T00:23:58Z [verbose] Readiness Indicator file check\\\\n2025-12-03T00:24:43Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m8pj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m2wz8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.632720 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"693d6bb6-3b42-408e-a6e2-43c9ac5fcda0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a30bdbb17b4bc29d1985cd2d49e1b8365983cf28e0d36b37c44ef0c04ef7455a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e86cacc8f8deada71ee3ce04595fbf26325281893c20be9aa672de27a67541f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zd86r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ng944\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.665810 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d5e7779-476a-4b6f-8153-7fd40fab8d90\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T00:24:53Z\\\",\\\"message\\\":\\\"t: *v1.Pod openshift-multus/network-metrics-daemon-kdznz\\\\nI1203 00:24:53.061488 6904 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 in node crc\\\\nI1203 00:24:53.061567 6904 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-kdznz in node crc\\\\nI1203 00:24:53.061573 6904 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944 after 0 failed attempt(s)\\\\nF1203 00:24:53.061578 6904 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:24:52Z is after 2025-08-24T17:21:41Z]\\\\nI1203 00:24:53.061\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hs5bf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f29xt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.689738 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-59mch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"36b8f8fb-72ad-407e-914c-9f71ada6e67d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16afecfd3dd4d7e7bea2958c488498fcf11de4591dd4197698d90d65363a3d4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:24:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f366e62dc0160ba835603823a1ae15b187cefa15c29f2da82e019601a105d9ac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03cbf855ffdf300cdc6538fa748640a08c6a5d24eae38cace8762df36e9a2403\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a12f300f4aafcde28ec92ed95492bc0627f181eb07691ed0733ef4eafaa4136b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e147a2002094fa56a84396b8c9fcf5b749ee3d6331873aa1e4302aa0d19800d3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc62cb264fcdee752f4148cbd75bb73cd068321c2af22e70355b2716f419347a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0ba85704b454c87b1f6e533bdf95cad26bb5c760e7ad9098d85d7c91818283f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:24:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:24:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cdrqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-59mch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.702077 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.702250 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.702266 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.702314 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.702329 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:14Z","lastTransitionTime":"2025-12-03T00:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.708933 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1e4151fa-f167-43cc-9ed0-4480bb85d1d5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5c5fdbc67a43f79f89cae267a16670b896ba2ac5c7099dd63f9be7800fe13302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c216b2bf5454b94499b8371ff9cfdc53bc95628ca8152b15c54d6d81751dc22\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b45acd848618f8e44429ede26117192520537bcf9c9f3a8cd08272266c920402\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.723713 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bef494-789f-4aa3-8e7a-0bf1b567f4c6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://593f8061c9279d7ce942487a0aadaf0863c84c501231d5d6371ea2925f6590f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0810f70f8be3d87c3806e90613947b90a33a756b8763dab196a36c0a1bf6797\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2cf017fae5258839967eff783a90be29e7ba41fd02d9077407c1c6fc20f05fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7707f16909775c0967b06a674ec1d84c62b32e6a466dace80ec27e9053669a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.740151 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.756190 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.770618 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92796847-a53f-4ffd-8746-7fd65c0ee4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0264849ef5cfda00c0794b79875938f5110b39103e025d6c8d96e0055d609077\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lbpvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2qk89\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.797831 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39b9d3e9-db75-4a63-88df-a2f397006156\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9710273872e1ef5c52f498af9fec43d363488c0d82f587434b5cd1b85c64d63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82e6b7840521ea6ab311f64e9dba441f6101e448117958e796259d68a0628dd4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a60a0ad4560c45828cb795a3024abcd0ce02b775965e6ac3061e78e5b6506c6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13a2f70a71a9db2893787a34d04a324f4eabf12f088a2bf28c9506dc7f412d7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ffada99477aaa54b98e1de927197ec844b7f7396ffe7f50b365e8bc6600f038\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f222a436af147711d5c22ff31aa526b70f4d28bc259710e79b4b3901d3e0a832\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fd1277a874e006421242bf8eda0769741159da24e15c12216d4ad20a5917b01\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2876e01c70ab08419be327bea11a2b13c7e1d8050019820cb4e0bf75df85689a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T00:23:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T00:23:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:34Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.805761 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.805827 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.805842 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.805889 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.805904 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:14Z","lastTransitionTime":"2025-12-03T00:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.815283 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://876a15fc066e744248ed63f93aebc8377bfc65d3bc2c189776019429c9702e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.830316 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://421d3adc2dfa990c90b73d591b9762d4bf2625c254cbb29b4ae1dfcf3af3baf6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.844957 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tnw7s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c92a06d2-780e-4372-9395-8519acdb616d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://234bccd998e9322cf548ebe641856373c8b01d1c4fc90d0483255d08d8992a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hkkqq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tnw7s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.857708 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kdznz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e49d5e0a-f59c-4dc1-9140-4b255576cd72\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hbfnp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:24:09Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kdznz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.875310 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.888513 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fe938c8df9bd10fd586e3e3ad44d1e140ec048f88d88560335c3a2502897f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3fe436971144b97277500d41a754750554de3efffe49b877322f9f9183120393\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.900965 4912 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wnn4k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"114ee369-a48e-4c87-9c8b-9436a230f02e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T00:23:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9304315129cf6fddcd23b0f651043851b404da2945d6e169af2e9cfa16b5461\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T00:23:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w6b2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T00:23:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wnn4k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:14Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.908010 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.908078 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.908095 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.908120 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:14 crc kubenswrapper[4912]: I1203 00:25:14.908138 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:14Z","lastTransitionTime":"2025-12-03T00:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.010554 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.010612 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.010629 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.010670 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.010690 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.113740 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.113802 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.113822 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.113847 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.113866 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.217269 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.217331 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.217351 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.217375 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.217393 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.319396 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.319496 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.319516 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.319543 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.319562 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.422035 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.422074 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.422085 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.422099 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.422110 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.524490 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.524908 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.525111 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.525275 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.525491 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.571147 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:15 crc kubenswrapper[4912]: E1203 00:25:15.571647 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.627496 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.627530 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.627539 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.627552 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.627562 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.730377 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.730576 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.730585 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.730605 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.730617 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.832820 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.832880 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.832890 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.832935 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.832947 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.934990 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.935031 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.935044 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.935061 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:15 crc kubenswrapper[4912]: I1203 00:25:15.935073 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:15Z","lastTransitionTime":"2025-12-03T00:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.037606 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.037647 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.037656 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.037671 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.037680 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.140218 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.140267 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.140279 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.140295 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.140308 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.243036 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.243084 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.243098 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.243117 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.243135 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.345738 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.345802 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.345821 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.345851 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.345870 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.448375 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.448735 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.448832 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.448938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.449033 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.551827 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.551890 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.551901 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.551921 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.551934 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.571161 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:16 crc kubenswrapper[4912]: E1203 00:25:16.571316 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.571181 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.571373 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:16 crc kubenswrapper[4912]: E1203 00:25:16.571474 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:16 crc kubenswrapper[4912]: E1203 00:25:16.571573 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.654852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.655453 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.655487 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.655504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.655514 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.758148 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.758189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.758200 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.758217 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.758227 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.860999 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.861106 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.861127 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.861152 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.861170 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.963085 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.963150 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.963171 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.963199 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:16 crc kubenswrapper[4912]: I1203 00:25:16.963220 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:16Z","lastTransitionTime":"2025-12-03T00:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.066202 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.066293 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.066313 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.066334 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.066350 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.168829 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.168893 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.168911 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.168934 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.168951 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.270809 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.270892 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.270917 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.270944 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.270961 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.373852 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.373993 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.374195 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.374223 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.374240 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.476521 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.476575 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.476625 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.476657 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.476674 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.571014 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:17 crc kubenswrapper[4912]: E1203 00:25:17.571364 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.578673 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.578710 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.578719 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.578732 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.578743 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.636155 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.636224 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.636243 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.636269 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.636332 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: E1203 00:25:17.649011 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.653116 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.653150 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.653160 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.653176 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.653188 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: E1203 00:25:17.664879 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.668524 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.668554 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.668565 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.668580 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.668590 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: E1203 00:25:17.680171 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.684127 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.684160 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.684169 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.684229 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.684241 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: E1203 00:25:17.697410 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.701650 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.701708 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.701724 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.701749 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.701770 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: E1203 00:25:17.715924 4912 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T00:25:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"bc17af26-6c16-455e-9b60-e30600d665e4\\\",\\\"systemUUID\\\":\\\"92184f41-123a-4ffc-8bda-9fa272ab6757\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T00:25:17Z is after 2025-08-24T17:21:41Z" Dec 03 00:25:17 crc kubenswrapper[4912]: E1203 00:25:17.716192 4912 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.718189 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.718227 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.718238 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.718255 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.718266 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.821504 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.821540 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.821549 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.821565 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.821575 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.924284 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.924353 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.924365 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.924384 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:17 crc kubenswrapper[4912]: I1203 00:25:17.924396 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:17Z","lastTransitionTime":"2025-12-03T00:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.026183 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.026225 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.026234 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.026248 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.026259 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.129056 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.129094 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.129103 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.129120 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.129132 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.232365 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.232414 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.232423 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.232451 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.232461 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.335205 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.335264 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.335274 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.335290 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.335300 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.445202 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.445254 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.445267 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.445285 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.445295 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.547309 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.547346 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.547355 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.547370 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.547380 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.571138 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:18 crc kubenswrapper[4912]: E1203 00:25:18.571237 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.571294 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:18 crc kubenswrapper[4912]: E1203 00:25:18.571497 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.572218 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:18 crc kubenswrapper[4912]: E1203 00:25:18.572557 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.572867 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:25:18 crc kubenswrapper[4912]: E1203 00:25:18.573159 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.649527 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.649601 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.649619 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.649645 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.649662 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.752183 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.752222 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.752231 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.752246 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.752256 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.854905 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.854974 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.854987 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.855008 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.855023 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.957264 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.957307 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.957316 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.957332 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:18 crc kubenswrapper[4912]: I1203 00:25:18.957348 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:18Z","lastTransitionTime":"2025-12-03T00:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.060933 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.060980 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.060992 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.061015 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.061034 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.163037 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.163086 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.163098 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.163115 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.163126 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.266036 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.266079 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.266089 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.266103 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.266112 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.369747 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.369808 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.369820 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.369837 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.369851 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.472580 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.472669 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.472697 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.472730 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.472756 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.571005 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:19 crc kubenswrapper[4912]: E1203 00:25:19.571184 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.575465 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.575520 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.575532 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.575547 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.575557 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.677581 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.677615 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.677623 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.677636 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.677645 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.780222 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.780270 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.780284 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.780306 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.780320 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.882960 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.883021 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.883042 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.883069 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.883088 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.985573 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.985599 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.985608 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.985619 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:19 crc kubenswrapper[4912]: I1203 00:25:19.985627 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:19Z","lastTransitionTime":"2025-12-03T00:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.087590 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.087625 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.087636 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.087652 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.087660 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:20Z","lastTransitionTime":"2025-12-03T00:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.190522 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.190560 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.190571 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.190589 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.190601 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:20Z","lastTransitionTime":"2025-12-03T00:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.292944 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.293037 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.293053 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.293068 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.293078 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:20Z","lastTransitionTime":"2025-12-03T00:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.395295 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.395336 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.395344 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.395361 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.395371 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:20Z","lastTransitionTime":"2025-12-03T00:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.497575 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.497608 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.497617 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.497634 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.497643 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:20Z","lastTransitionTime":"2025-12-03T00:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.571353 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.571354 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.571485 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:20 crc kubenswrapper[4912]: E1203 00:25:20.571555 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:20 crc kubenswrapper[4912]: E1203 00:25:20.571769 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:20 crc kubenswrapper[4912]: E1203 00:25:20.571839 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.599692 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.599730 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.599739 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.599754 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.599764 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:20Z","lastTransitionTime":"2025-12-03T00:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.703215 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.703296 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.703322 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.703353 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.703377 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:20Z","lastTransitionTime":"2025-12-03T00:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.805729 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.805773 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.805785 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.805803 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.805815 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:20Z","lastTransitionTime":"2025-12-03T00:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.908233 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.908321 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.908335 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.908351 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:20 crc kubenswrapper[4912]: I1203 00:25:20.908364 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:20Z","lastTransitionTime":"2025-12-03T00:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.011177 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.011243 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.011261 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.011290 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.011313 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.113828 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.113874 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.113882 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.113896 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.113905 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.215927 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.215969 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.215980 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.215995 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.216007 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.317996 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.318043 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.318056 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.318083 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.318098 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.420613 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.420652 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.420685 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.420703 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.420713 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.523304 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.523344 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.523381 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.523400 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.523411 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.570946 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:21 crc kubenswrapper[4912]: E1203 00:25:21.571157 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.625733 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.625771 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.625779 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.625792 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.625802 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.730995 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.731030 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.731040 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.731053 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.731063 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.834232 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.834287 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.834300 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.834317 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.834329 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.936781 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.936825 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.936836 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.936853 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:21 crc kubenswrapper[4912]: I1203 00:25:21.936865 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:21Z","lastTransitionTime":"2025-12-03T00:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.040325 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.040369 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.040379 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.040393 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.040403 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.142809 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.142872 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.142885 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.142904 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.142920 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.245747 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.245816 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.245829 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.245847 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.245859 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.348447 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.348567 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.348580 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.348596 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.348607 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.452401 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.452524 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.452551 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.452582 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.452645 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.555125 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.555186 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.555205 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.555230 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.555256 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.571561 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.571711 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.571774 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:22 crc kubenswrapper[4912]: E1203 00:25:22.571744 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:22 crc kubenswrapper[4912]: E1203 00:25:22.572013 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:22 crc kubenswrapper[4912]: E1203 00:25:22.572261 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.658220 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.658292 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.658310 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.658340 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.658360 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.761278 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.761351 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.761362 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.761376 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.761406 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.864996 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.865061 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.865083 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.865112 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.865135 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.968165 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.968246 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.968271 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.968298 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:22 crc kubenswrapper[4912]: I1203 00:25:22.968313 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:22Z","lastTransitionTime":"2025-12-03T00:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.072696 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.072748 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.072759 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.072778 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.072793 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.176299 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.176351 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.176362 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.176384 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.176399 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.278418 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.278503 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.278526 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.278544 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.278557 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.381204 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.381244 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.381254 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.381268 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.381278 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.484382 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.484440 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.484454 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.484469 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.484479 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.571521 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:23 crc kubenswrapper[4912]: E1203 00:25:23.571653 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.586094 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.586143 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.586155 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.586172 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.586186 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.688837 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.688878 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.688888 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.688902 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.688911 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.791609 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.791656 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.791664 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.791680 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.791688 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.894486 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.894540 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.894549 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.894568 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.894579 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.996938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.996986 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.996995 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.997008 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:23 crc kubenswrapper[4912]: I1203 00:25:23.997017 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:23Z","lastTransitionTime":"2025-12-03T00:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.099006 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.099053 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.099064 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.099081 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.099093 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:24Z","lastTransitionTime":"2025-12-03T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.201594 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.201636 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.201647 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.201663 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.201673 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:24Z","lastTransitionTime":"2025-12-03T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.303684 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.303721 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.303733 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.303749 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.303762 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:24Z","lastTransitionTime":"2025-12-03T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.405738 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.405787 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.405799 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.405816 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.405826 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:24Z","lastTransitionTime":"2025-12-03T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.510722 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.510790 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.510802 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.510822 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.510839 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:24Z","lastTransitionTime":"2025-12-03T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.571177 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:24 crc kubenswrapper[4912]: E1203 00:25:24.571321 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.571371 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.571572 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:24 crc kubenswrapper[4912]: E1203 00:25:24.571570 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:24 crc kubenswrapper[4912]: E1203 00:25:24.571660 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.600460 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ng944" podStartSLOduration=90.600408778 podStartE2EDuration="1m30.600408778s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.590162362 +0000 UTC m=+110.232182942" watchObservedRunningTime="2025-12-03 00:25:24.600408778 +0000 UTC m=+110.242429348" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.600875 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=36.600861612 podStartE2EDuration="36.600861612s" podCreationTimestamp="2025-12-03 00:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.600751919 +0000 UTC m=+110.242772519" watchObservedRunningTime="2025-12-03 00:25:24.600861612 +0000 UTC m=+110.242882192" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.613275 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.613321 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.613333 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.613350 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.613362 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:24Z","lastTransitionTime":"2025-12-03T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.632633 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=92.632614469 podStartE2EDuration="1m32.632614469s" podCreationTimestamp="2025-12-03 00:23:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.617598666 +0000 UTC m=+110.259619256" watchObservedRunningTime="2025-12-03 00:25:24.632614469 +0000 UTC m=+110.274635029" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.647629 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-m2wz8" podStartSLOduration=90.647610203 podStartE2EDuration="1m30.647610203s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.632496946 +0000 UTC m=+110.274517536" watchObservedRunningTime="2025-12-03 00:25:24.647610203 +0000 UTC m=+110.289630763" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.662341 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podStartSLOduration=90.662325739 podStartE2EDuration="1m30.662325739s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.662077071 +0000 UTC m=+110.304097651" watchObservedRunningTime="2025-12-03 00:25:24.662325739 +0000 UTC m=+110.304346299" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.709331 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-59mch" podStartSLOduration=90.709309977 podStartE2EDuration="1m30.709309977s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.697897637 +0000 UTC m=+110.339918207" watchObservedRunningTime="2025-12-03 00:25:24.709309977 +0000 UTC m=+110.351330537" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.709521 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=92.709515822 podStartE2EDuration="1m32.709515822s" podCreationTimestamp="2025-12-03 00:23:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.708968187 +0000 UTC m=+110.350988757" watchObservedRunningTime="2025-12-03 00:25:24.709515822 +0000 UTC m=+110.351536382" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.714995 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.715037 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.715046 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.715059 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.715068 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:24Z","lastTransitionTime":"2025-12-03T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.720697 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=58.720683895 podStartE2EDuration="58.720683895s" podCreationTimestamp="2025-12-03 00:24:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.720683165 +0000 UTC m=+110.362703735" watchObservedRunningTime="2025-12-03 00:25:24.720683895 +0000 UTC m=+110.362704455" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.740704 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-tnw7s" podStartSLOduration=91.740685815 podStartE2EDuration="1m31.740685815s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.740268382 +0000 UTC m=+110.382288942" watchObservedRunningTime="2025-12-03 00:25:24.740685815 +0000 UTC m=+110.382706375" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.770571 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=91.770549688 podStartE2EDuration="1m31.770549688s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.769464216 +0000 UTC m=+110.411484796" watchObservedRunningTime="2025-12-03 00:25:24.770549688 +0000 UTC m=+110.412570238" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.816958 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.816998 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.817010 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.817025 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.817036 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:24Z","lastTransitionTime":"2025-12-03T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.859672 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-wnn4k" podStartSLOduration=91.859652783 podStartE2EDuration="1m31.859652783s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:24.859129288 +0000 UTC m=+110.501149858" watchObservedRunningTime="2025-12-03 00:25:24.859652783 +0000 UTC m=+110.501673353" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.919162 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.919736 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.919817 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.919894 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:24 crc kubenswrapper[4912]: I1203 00:25:24.919971 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:24Z","lastTransitionTime":"2025-12-03T00:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.022346 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.022390 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.022402 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.022452 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.022470 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.127390 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.127768 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.127887 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.128010 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.128110 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.231250 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.231283 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.231306 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.231319 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.231327 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.333562 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.333609 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.333626 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.333648 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.333667 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.435845 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.435885 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.435895 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.435910 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.435922 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.538688 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.538720 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.538728 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.538763 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.538779 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.571047 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:25 crc kubenswrapper[4912]: E1203 00:25:25.571160 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.640672 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.640734 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.640745 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.640776 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.640789 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.743513 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.743566 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.743581 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.743600 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.743614 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.845636 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.845688 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.845699 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.845716 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.845727 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.948624 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.948668 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.948685 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.948703 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:25 crc kubenswrapper[4912]: I1203 00:25:25.948715 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:25Z","lastTransitionTime":"2025-12-03T00:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.051192 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.051247 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.051259 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.051277 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.051289 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.153938 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.153978 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.153986 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.153999 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.154009 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.256502 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.256550 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.256561 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.256578 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.256589 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.358701 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.358735 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.358744 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.358761 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.358772 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.461082 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.461119 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.461129 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.461145 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.461158 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.563541 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.563601 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.563613 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.563627 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.563639 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.570879 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.570898 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:26 crc kubenswrapper[4912]: E1203 00:25:26.570986 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.571015 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:26 crc kubenswrapper[4912]: E1203 00:25:26.571082 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:26 crc kubenswrapper[4912]: E1203 00:25:26.571132 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.667255 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.667289 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.667302 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.667315 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.667324 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.770198 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.770275 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.770299 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.770332 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.770361 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.872774 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.872827 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.872849 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.872879 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.872901 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.975608 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.975665 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.975677 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.975695 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:26 crc kubenswrapper[4912]: I1203 00:25:26.975707 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:26Z","lastTransitionTime":"2025-12-03T00:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.078250 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.078307 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.078325 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.078340 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.078350 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:27Z","lastTransitionTime":"2025-12-03T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.180263 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.180297 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.180306 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.180319 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.180327 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:27Z","lastTransitionTime":"2025-12-03T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.282844 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.282877 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.282904 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.282918 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.282926 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:27Z","lastTransitionTime":"2025-12-03T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.385455 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.385497 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.385506 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.385519 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.385528 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:27Z","lastTransitionTime":"2025-12-03T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.488042 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.488073 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.488081 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.488095 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.488103 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:27Z","lastTransitionTime":"2025-12-03T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.570569 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:27 crc kubenswrapper[4912]: E1203 00:25:27.570689 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.590534 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.590577 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.590587 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.590600 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.590611 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:27Z","lastTransitionTime":"2025-12-03T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.692464 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.692505 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.692515 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.692528 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.692539 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:27Z","lastTransitionTime":"2025-12-03T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.794996 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.795030 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.795062 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.795075 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.795086 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:27Z","lastTransitionTime":"2025-12-03T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.830952 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.831030 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.831043 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.831058 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.831068 4912 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T00:25:27Z","lastTransitionTime":"2025-12-03T00:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.867252 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh"] Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.867658 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.869882 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.869945 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.869992 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.869948 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.960600 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a1bc3afa-7b58-410e-9029-6b68a439affe-service-ca\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.960746 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a1bc3afa-7b58-410e-9029-6b68a439affe-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.960798 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1bc3afa-7b58-410e-9029-6b68a439affe-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.960836 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a1bc3afa-7b58-410e-9029-6b68a439affe-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:27 crc kubenswrapper[4912]: I1203 00:25:27.960906 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a1bc3afa-7b58-410e-9029-6b68a439affe-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.061706 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a1bc3afa-7b58-410e-9029-6b68a439affe-service-ca\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.061797 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a1bc3afa-7b58-410e-9029-6b68a439affe-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.061844 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1bc3afa-7b58-410e-9029-6b68a439affe-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.061890 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a1bc3afa-7b58-410e-9029-6b68a439affe-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.061896 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a1bc3afa-7b58-410e-9029-6b68a439affe-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.061949 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a1bc3afa-7b58-410e-9029-6b68a439affe-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.062031 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a1bc3afa-7b58-410e-9029-6b68a439affe-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.062676 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a1bc3afa-7b58-410e-9029-6b68a439affe-service-ca\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.067676 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1bc3afa-7b58-410e-9029-6b68a439affe-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.078824 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a1bc3afa-7b58-410e-9029-6b68a439affe-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-cx5mh\" (UID: \"a1bc3afa-7b58-410e-9029-6b68a439affe\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.185716 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.571290 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:28 crc kubenswrapper[4912]: E1203 00:25:28.572229 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.571584 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:28 crc kubenswrapper[4912]: I1203 00:25:28.571571 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:28 crc kubenswrapper[4912]: E1203 00:25:28.572420 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:28 crc kubenswrapper[4912]: E1203 00:25:28.572690 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:29 crc kubenswrapper[4912]: I1203 00:25:29.102025 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" event={"ID":"a1bc3afa-7b58-410e-9029-6b68a439affe","Type":"ContainerStarted","Data":"beb10ee7554e90bf79cdeacc17a36a0a66b5ab5a4a0c3540539e5f9d4b5c6387"} Dec 03 00:25:29 crc kubenswrapper[4912]: I1203 00:25:29.102076 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" event={"ID":"a1bc3afa-7b58-410e-9029-6b68a439affe","Type":"ContainerStarted","Data":"32dd256c0e644ecfcfb5cde00edf9c59d39d46e868beaf036bb3baae0302bf3b"} Dec 03 00:25:29 crc kubenswrapper[4912]: I1203 00:25:29.115550 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-cx5mh" podStartSLOduration=95.115529004 podStartE2EDuration="1m35.115529004s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:29.11436205 +0000 UTC m=+114.756382640" watchObservedRunningTime="2025-12-03 00:25:29.115529004 +0000 UTC m=+114.757549564" Dec 03 00:25:29 crc kubenswrapper[4912]: I1203 00:25:29.570968 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:29 crc kubenswrapper[4912]: E1203 00:25:29.571315 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.106198 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2wz8_8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77/kube-multus/1.log" Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.106692 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2wz8_8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77/kube-multus/0.log" Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.106729 4912 generic.go:334] "Generic (PLEG): container finished" podID="8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77" containerID="62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e" exitCode=1 Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.106756 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2wz8" event={"ID":"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77","Type":"ContainerDied","Data":"62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e"} Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.106787 4912 scope.go:117] "RemoveContainer" containerID="d0b1207a0204e9fdd345f45c97eeb8edf376376e55ec41a77ffa32feae05aff6" Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.107181 4912 scope.go:117] "RemoveContainer" containerID="62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e" Dec 03 00:25:30 crc kubenswrapper[4912]: E1203 00:25:30.107365 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-m2wz8_openshift-multus(8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77)\"" pod="openshift-multus/multus-m2wz8" podUID="8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77" Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.571667 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.571699 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.571693 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:30 crc kubenswrapper[4912]: E1203 00:25:30.571790 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:30 crc kubenswrapper[4912]: E1203 00:25:30.571957 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:30 crc kubenswrapper[4912]: E1203 00:25:30.572017 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:30 crc kubenswrapper[4912]: I1203 00:25:30.573577 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:25:30 crc kubenswrapper[4912]: E1203 00:25:30.574144 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f29xt_openshift-ovn-kubernetes(7d5e7779-476a-4b6f-8153-7fd40fab8d90)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" Dec 03 00:25:31 crc kubenswrapper[4912]: I1203 00:25:31.570849 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:31 crc kubenswrapper[4912]: E1203 00:25:31.571232 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:32 crc kubenswrapper[4912]: I1203 00:25:32.112696 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2wz8_8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77/kube-multus/1.log" Dec 03 00:25:32 crc kubenswrapper[4912]: I1203 00:25:32.570586 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:32 crc kubenswrapper[4912]: I1203 00:25:32.570669 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:32 crc kubenswrapper[4912]: I1203 00:25:32.570719 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:32 crc kubenswrapper[4912]: E1203 00:25:32.570781 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:32 crc kubenswrapper[4912]: E1203 00:25:32.570878 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:32 crc kubenswrapper[4912]: E1203 00:25:32.570977 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:33 crc kubenswrapper[4912]: I1203 00:25:33.571392 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:33 crc kubenswrapper[4912]: E1203 00:25:33.571553 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:34 crc kubenswrapper[4912]: E1203 00:25:34.562546 4912 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 00:25:34 crc kubenswrapper[4912]: I1203 00:25:34.571817 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:34 crc kubenswrapper[4912]: I1203 00:25:34.571747 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:34 crc kubenswrapper[4912]: E1203 00:25:34.576396 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:34 crc kubenswrapper[4912]: I1203 00:25:34.576459 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:34 crc kubenswrapper[4912]: E1203 00:25:34.576590 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:34 crc kubenswrapper[4912]: E1203 00:25:34.576696 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:34 crc kubenswrapper[4912]: E1203 00:25:34.657223 4912 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:25:35 crc kubenswrapper[4912]: I1203 00:25:35.570883 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:35 crc kubenswrapper[4912]: E1203 00:25:35.571269 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:36 crc kubenswrapper[4912]: I1203 00:25:36.570547 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:36 crc kubenswrapper[4912]: I1203 00:25:36.570645 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:36 crc kubenswrapper[4912]: E1203 00:25:36.570700 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:36 crc kubenswrapper[4912]: I1203 00:25:36.570732 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:36 crc kubenswrapper[4912]: E1203 00:25:36.570844 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:36 crc kubenswrapper[4912]: E1203 00:25:36.570913 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:37 crc kubenswrapper[4912]: I1203 00:25:37.571453 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:37 crc kubenswrapper[4912]: E1203 00:25:37.571577 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:38 crc kubenswrapper[4912]: I1203 00:25:38.571014 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:38 crc kubenswrapper[4912]: I1203 00:25:38.571186 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:38 crc kubenswrapper[4912]: E1203 00:25:38.571288 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:38 crc kubenswrapper[4912]: I1203 00:25:38.571362 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:38 crc kubenswrapper[4912]: E1203 00:25:38.571453 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:38 crc kubenswrapper[4912]: E1203 00:25:38.571570 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:39 crc kubenswrapper[4912]: I1203 00:25:39.571477 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:39 crc kubenswrapper[4912]: E1203 00:25:39.571874 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:39 crc kubenswrapper[4912]: E1203 00:25:39.658228 4912 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:25:40 crc kubenswrapper[4912]: I1203 00:25:40.570962 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:40 crc kubenswrapper[4912]: E1203 00:25:40.571090 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:40 crc kubenswrapper[4912]: I1203 00:25:40.571300 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:40 crc kubenswrapper[4912]: E1203 00:25:40.571399 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:40 crc kubenswrapper[4912]: I1203 00:25:40.571699 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:40 crc kubenswrapper[4912]: E1203 00:25:40.571780 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:41 crc kubenswrapper[4912]: I1203 00:25:41.570965 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:41 crc kubenswrapper[4912]: E1203 00:25:41.571105 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:41 crc kubenswrapper[4912]: I1203 00:25:41.571882 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:25:42 crc kubenswrapper[4912]: I1203 00:25:42.570894 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:42 crc kubenswrapper[4912]: I1203 00:25:42.570932 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:42 crc kubenswrapper[4912]: I1203 00:25:42.570894 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:42 crc kubenswrapper[4912]: E1203 00:25:42.571036 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:42 crc kubenswrapper[4912]: E1203 00:25:42.571084 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:42 crc kubenswrapper[4912]: E1203 00:25:42.571211 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:43 crc kubenswrapper[4912]: I1203 00:25:43.145384 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/3.log" Dec 03 00:25:43 crc kubenswrapper[4912]: I1203 00:25:43.148028 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerStarted","Data":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} Dec 03 00:25:43 crc kubenswrapper[4912]: I1203 00:25:43.149109 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:25:43 crc kubenswrapper[4912]: I1203 00:25:43.176744 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podStartSLOduration=109.176728904 podStartE2EDuration="1m49.176728904s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:25:43.175673814 +0000 UTC m=+128.817694394" watchObservedRunningTime="2025-12-03 00:25:43.176728904 +0000 UTC m=+128.818749464" Dec 03 00:25:43 crc kubenswrapper[4912]: I1203 00:25:43.323058 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kdznz"] Dec 03 00:25:43 crc kubenswrapper[4912]: I1203 00:25:43.323179 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:43 crc kubenswrapper[4912]: E1203 00:25:43.323261 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:43 crc kubenswrapper[4912]: I1203 00:25:43.571157 4912 scope.go:117] "RemoveContainer" containerID="62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e" Dec 03 00:25:44 crc kubenswrapper[4912]: I1203 00:25:44.570685 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:44 crc kubenswrapper[4912]: I1203 00:25:44.570733 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:44 crc kubenswrapper[4912]: I1203 00:25:44.570752 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:44 crc kubenswrapper[4912]: I1203 00:25:44.570795 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:44 crc kubenswrapper[4912]: E1203 00:25:44.573264 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:44 crc kubenswrapper[4912]: E1203 00:25:44.573420 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:44 crc kubenswrapper[4912]: E1203 00:25:44.573643 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:44 crc kubenswrapper[4912]: E1203 00:25:44.573801 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:44 crc kubenswrapper[4912]: E1203 00:25:44.658729 4912 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:25:45 crc kubenswrapper[4912]: I1203 00:25:45.155591 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2wz8_8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77/kube-multus/1.log" Dec 03 00:25:45 crc kubenswrapper[4912]: I1203 00:25:45.155638 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2wz8" event={"ID":"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77","Type":"ContainerStarted","Data":"892e7bbdead6a107c57ae944c2e00a9f798c609b0c8b755fb63c17c64a088a49"} Dec 03 00:25:46 crc kubenswrapper[4912]: I1203 00:25:46.571479 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:46 crc kubenswrapper[4912]: I1203 00:25:46.571517 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:46 crc kubenswrapper[4912]: I1203 00:25:46.571580 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:46 crc kubenswrapper[4912]: E1203 00:25:46.571614 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:46 crc kubenswrapper[4912]: I1203 00:25:46.571649 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:46 crc kubenswrapper[4912]: E1203 00:25:46.571737 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:46 crc kubenswrapper[4912]: E1203 00:25:46.571832 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:46 crc kubenswrapper[4912]: E1203 00:25:46.572037 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:48 crc kubenswrapper[4912]: I1203 00:25:48.570801 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:48 crc kubenswrapper[4912]: I1203 00:25:48.570883 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:48 crc kubenswrapper[4912]: I1203 00:25:48.571003 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:48 crc kubenswrapper[4912]: E1203 00:25:48.571022 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 00:25:48 crc kubenswrapper[4912]: I1203 00:25:48.571078 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:48 crc kubenswrapper[4912]: E1203 00:25:48.571113 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kdznz" podUID="e49d5e0a-f59c-4dc1-9140-4b255576cd72" Dec 03 00:25:48 crc kubenswrapper[4912]: E1203 00:25:48.571267 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 00:25:48 crc kubenswrapper[4912]: E1203 00:25:48.571394 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.571762 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.571857 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.571769 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.571755 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.574573 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.574609 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.575028 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.575112 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.575237 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 00:25:50 crc kubenswrapper[4912]: I1203 00:25:50.575902 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 00:25:54 crc kubenswrapper[4912]: I1203 00:25:54.887241 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.860514 4912 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.892653 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tqgnx"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.893183 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.893822 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.894690 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.894876 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lg5fq"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.895662 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.896135 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.896514 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.896592 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.896842 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.897487 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.900676 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.902261 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.907034 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.907148 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.907195 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.907343 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.907449 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.907492 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908039 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908305 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908369 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908466 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908500 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908693 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908736 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908847 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908899 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908952 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.908903 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.909154 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.912871 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.912909 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.912958 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.912874 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.912985 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.914706 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.914727 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.914754 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.915922 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.916461 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.928080 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.928694 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.954951 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.955243 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7x9ms"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.955273 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.955350 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.955771 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.955903 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.956220 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.956491 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.956847 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.956972 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.957423 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.957662 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.958673 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-b5vr9"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.959088 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b5vr9" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.959820 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.962792 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7ckmb"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.963315 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.964061 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29412000-5584k"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.974648 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.975827 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.977869 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.982088 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kd5n5"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.983171 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.996930 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-c65h9"] Dec 03 00:25:58 crc kubenswrapper[4912]: I1203 00:25:58.997620 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.001950 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.001805 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.002144 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.003921 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.005092 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.009305 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.009479 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.009305 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.009913 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.010177 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.010255 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.016741 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.018298 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.018599 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.018854 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.021444 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.021886 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.028984 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.029383 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.029395 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.029663 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.030141 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.031289 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-cg7ls"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.031842 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.032334 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.032743 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.037291 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-dddxf"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.038015 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.045010 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.045277 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.045365 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.045490 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.045666 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.045923 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.046073 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.046312 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.046379 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.046544 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.046626 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.046661 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.046902 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.047031 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4pwws"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.047194 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.047391 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.047592 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.047718 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.047756 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.047907 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.048059 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.048240 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.048276 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.048338 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.048383 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.048397 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.048576 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.048728 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.049212 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.050368 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.051360 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.051584 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.051669 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.051934 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.052150 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.052243 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.052402 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.052497 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.052651 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.052695 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.052747 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.052840 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.052941 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.053065 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.053146 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.053256 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbppn"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.053856 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.054186 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.054219 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.054377 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.054820 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.055295 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.055998 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.056803 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.057997 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.077836 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.080809 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.082004 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.085741 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.106804 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.107587 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.107629 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.108208 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.108389 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tqgnx"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.108726 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109366 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109393 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fffn6\" (UniqueName: \"kubernetes.io/projected/975f5428-8720-4c72-8595-f213ef97888e-kube-api-access-fffn6\") pod \"openshift-config-operator-7777fb866f-c65h9\" (UID: \"975f5428-8720-4c72-8595-f213ef97888e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109414 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109550 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-etcd-client\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109601 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j89rb\" (UniqueName: \"kubernetes.io/projected/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-kube-api-access-j89rb\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109630 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109704 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-client-ca\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109729 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-metrics-tls\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109825 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-serving-cert\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109926 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc341ad8-861e-427c-835c-b7bd7367b4a0-config\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.109954 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50003e5f-f76f-486a-bb7d-70a4e553e0cf-config\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.110034 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-audit-dir\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111524 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2839435a-a464-473c-94ad-7673ad00630b-serviceca\") pod \"image-pruner-29412000-5584k\" (UID: \"2839435a-a464-473c-94ad-7673ad00630b\") " pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111582 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/faa0cccb-7352-4752-8957-0d770dacaee5-audit-dir\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111601 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m76x\" (UniqueName: \"kubernetes.io/projected/61f377a2-06fc-427d-9c3e-38f9b3473e23-kube-api-access-7m76x\") pod \"downloads-7954f5f757-b5vr9\" (UID: \"61f377a2-06fc-427d-9c3e-38f9b3473e23\") " pod="openshift-console/downloads-7954f5f757-b5vr9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111641 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-serving-cert\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111663 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50003e5f-f76f-486a-bb7d-70a4e553e0cf-serving-cert\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111679 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/50003e5f-f76f-486a-bb7d-70a4e553e0cf-etcd-ca\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111696 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-encryption-config\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111728 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-config\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111748 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/fc341ad8-861e-427c-835c-b7bd7367b4a0-machine-approver-tls\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111770 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-audit\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111799 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-image-import-ca\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111814 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-config\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111830 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-audit-policies\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111846 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-encryption-config\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111879 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-etcd-serving-ca\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111896 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59f8296f-ab2b-40bc-af9c-80a1102321fa-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zzz2x\" (UID: \"59f8296f-ab2b-40bc-af9c-80a1102321fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111912 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50003e5f-f76f-486a-bb7d-70a4e553e0cf-etcd-client\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111961 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/975f5428-8720-4c72-8595-f213ef97888e-serving-cert\") pod \"openshift-config-operator-7777fb866f-c65h9\" (UID: \"975f5428-8720-4c72-8595-f213ef97888e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.111976 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-audit-dir\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112000 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-config\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112031 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112047 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk2q6\" (UniqueName: \"kubernetes.io/projected/fc341ad8-861e-427c-835c-b7bd7367b4a0-kube-api-access-lk2q6\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112065 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112081 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112111 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112131 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsvhn\" (UniqueName: \"kubernetes.io/projected/ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed-kube-api-access-fsvhn\") pod \"openshift-apiserver-operator-796bbdcf4f-zmv99\" (UID: \"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112154 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqrpv\" (UniqueName: \"kubernetes.io/projected/59f8296f-ab2b-40bc-af9c-80a1102321fa-kube-api-access-bqrpv\") pod \"machine-config-controller-84d6567774-zzz2x\" (UID: \"59f8296f-ab2b-40bc-af9c-80a1102321fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112183 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112199 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxm72\" (UniqueName: \"kubernetes.io/projected/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-kube-api-access-wxm72\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112213 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-846fp\" (UniqueName: \"kubernetes.io/projected/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-kube-api-access-846fp\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112229 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wptw\" (UniqueName: \"kubernetes.io/projected/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-kube-api-access-4wptw\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112259 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112286 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zmv99\" (UID: \"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112304 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/975f5428-8720-4c72-8595-f213ef97888e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-c65h9\" (UID: \"975f5428-8720-4c72-8595-f213ef97888e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112338 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d00bf964-db4a-449d-ad97-d563b4b70f50-serving-cert\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112353 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-client-ca\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112379 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112409 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59f8296f-ab2b-40bc-af9c-80a1102321fa-proxy-tls\") pod \"machine-config-controller-84d6567774-zzz2x\" (UID: \"59f8296f-ab2b-40bc-af9c-80a1102321fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112448 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112470 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g9v6\" (UniqueName: \"kubernetes.io/projected/faa0cccb-7352-4752-8957-0d770dacaee5-kube-api-access-4g9v6\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112490 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-node-pullsecrets\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112510 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112546 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112561 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zmv99\" (UID: \"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112576 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt7ft\" (UniqueName: \"kubernetes.io/projected/2839435a-a464-473c-94ad-7673ad00630b-kube-api-access-zt7ft\") pod \"image-pruner-29412000-5584k\" (UID: \"2839435a-a464-473c-94ad-7673ad00630b\") " pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112590 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-etcd-client\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112631 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112647 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112662 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qp8j\" (UniqueName: \"kubernetes.io/projected/f33e67e7-af60-4c02-86f0-1bfd3de3d690-kube-api-access-7qp8j\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112696 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgkwc\" (UniqueName: \"kubernetes.io/projected/50003e5f-f76f-486a-bb7d-70a4e553e0cf-kube-api-access-lgkwc\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112712 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-trusted-ca\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112726 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112741 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-images\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112771 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112947 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.112773 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g9n9\" (UniqueName: \"kubernetes.io/projected/d00bf964-db4a-449d-ad97-d563b4b70f50-kube-api-access-6g9n9\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.113112 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-config\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.113139 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.113156 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-audit-policies\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.113170 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fc341ad8-861e-427c-835c-b7bd7367b4a0-auth-proxy-config\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.113183 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/50003e5f-f76f-486a-bb7d-70a4e553e0cf-etcd-service-ca\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.113200 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f33e67e7-af60-4c02-86f0-1bfd3de3d690-serving-cert\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.113528 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.113769 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mv4s4"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.113908 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.114215 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t69lf"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.114741 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.116733 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.119021 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-njltp"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.119940 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-njltp" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.120376 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.126584 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.127391 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.129085 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4z6ww"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.129820 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.130046 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-bbm58"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.130415 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.131828 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgskz"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.132313 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.133541 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.134541 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.135054 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.135242 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.135939 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.135421 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.138776 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.139358 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.139376 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.139698 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b5vr9"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.139773 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.140082 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.140753 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7x9ms"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.142522 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lg5fq"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.143010 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-gqt86"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.143635 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-gqt86" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.145083 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.147034 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.148250 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.149574 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7ckmb"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.150782 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29412000-5584k"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.157928 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.159485 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.161214 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.165501 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.173293 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-n8lb6"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.174140 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.174579 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.176547 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-cg7ls"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.179513 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kd5n5"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.181620 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.183344 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbppn"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.187578 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.188967 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.191904 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.194743 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4pwws"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.195410 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.196511 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.197810 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.200867 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mv4s4"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.202055 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.203419 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-c65h9"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.205043 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgskz"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.206446 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4z6ww"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.208336 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.209325 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.210590 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.211964 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-gqt86"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.213763 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t69lf"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214058 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-node-pullsecrets\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214090 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214115 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214133 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zmv99\" (UID: \"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214155 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt7ft\" (UniqueName: \"kubernetes.io/projected/2839435a-a464-473c-94ad-7673ad00630b-kube-api-access-zt7ft\") pod \"image-pruner-29412000-5584k\" (UID: \"2839435a-a464-473c-94ad-7673ad00630b\") " pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214170 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-etcd-client\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214186 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214202 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214219 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qp8j\" (UniqueName: \"kubernetes.io/projected/f33e67e7-af60-4c02-86f0-1bfd3de3d690-kube-api-access-7qp8j\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214240 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc0875cd-1663-45f8-a70b-96b0fb672d08-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6cgnd\" (UID: \"dc0875cd-1663-45f8-a70b-96b0fb672d08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214256 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214273 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-trusted-ca-bundle\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214292 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgkwc\" (UniqueName: \"kubernetes.io/projected/50003e5f-f76f-486a-bb7d-70a4e553e0cf-kube-api-access-lgkwc\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214309 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-trusted-ca\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214323 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214340 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94036057-869a-49f0-84ca-8397bab8472c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xf9jw\" (UID: \"94036057-869a-49f0-84ca-8397bab8472c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214357 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-service-ca-bundle\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214374 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwh7l\" (UniqueName: \"kubernetes.io/projected/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-kube-api-access-fwh7l\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214389 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-oauth-serving-cert\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214405 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-images\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214440 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g9n9\" (UniqueName: \"kubernetes.io/projected/d00bf964-db4a-449d-ad97-d563b4b70f50-kube-api-access-6g9n9\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214463 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-config\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214482 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/23523b1a-5f26-4f73-8ae9-a9735a83a7de-srv-cert\") pod \"olm-operator-6b444d44fb-825zz\" (UID: \"23523b1a-5f26-4f73-8ae9-a9735a83a7de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214503 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214518 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-audit-policies\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214532 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fc341ad8-861e-427c-835c-b7bd7367b4a0-auth-proxy-config\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214548 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/50003e5f-f76f-486a-bb7d-70a4e553e0cf-etcd-service-ca\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214562 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f33e67e7-af60-4c02-86f0-1bfd3de3d690-serving-cert\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214579 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4hvq\" (UniqueName: \"kubernetes.io/projected/23523b1a-5f26-4f73-8ae9-a9735a83a7de-kube-api-access-w4hvq\") pod \"olm-operator-6b444d44fb-825zz\" (UID: \"23523b1a-5f26-4f73-8ae9-a9735a83a7de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214620 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214636 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fffn6\" (UniqueName: \"kubernetes.io/projected/975f5428-8720-4c72-8595-f213ef97888e-kube-api-access-fffn6\") pod \"openshift-config-operator-7777fb866f-c65h9\" (UID: \"975f5428-8720-4c72-8595-f213ef97888e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214654 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d9mkw\" (UID: \"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214669 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-serving-cert\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214686 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214702 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84edc937-d48f-439b-90a7-df4917ead6eb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5rhpw\" (UID: \"84edc937-d48f-439b-90a7-df4917ead6eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214718 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-images\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214755 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-etcd-client\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214772 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j89rb\" (UniqueName: \"kubernetes.io/projected/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-kube-api-access-j89rb\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214787 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214803 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-client-ca\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214820 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-metrics-tls\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214835 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-proxy-tls\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214851 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-serving-cert\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214867 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc341ad8-861e-427c-835c-b7bd7367b4a0-config\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214895 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50003e5f-f76f-486a-bb7d-70a4e553e0cf-config\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214911 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-audit-dir\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214928 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2839435a-a464-473c-94ad-7673ad00630b-serviceca\") pod \"image-pruner-29412000-5584k\" (UID: \"2839435a-a464-473c-94ad-7673ad00630b\") " pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214946 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94036057-869a-49f0-84ca-8397bab8472c-config\") pod \"kube-apiserver-operator-766d6c64bb-xf9jw\" (UID: \"94036057-869a-49f0-84ca-8397bab8472c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214962 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-metrics-certs\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214978 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77zl8\" (UniqueName: \"kubernetes.io/projected/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-kube-api-access-77zl8\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.214994 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/faa0cccb-7352-4752-8957-0d770dacaee5-audit-dir\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215015 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m76x\" (UniqueName: \"kubernetes.io/projected/61f377a2-06fc-427d-9c3e-38f9b3473e23-kube-api-access-7m76x\") pod \"downloads-7954f5f757-b5vr9\" (UID: \"61f377a2-06fc-427d-9c3e-38f9b3473e23\") " pod="openshift-console/downloads-7954f5f757-b5vr9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215036 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-serving-cert\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215054 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzc92\" (UniqueName: \"kubernetes.io/projected/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-kube-api-access-bzc92\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215073 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50003e5f-f76f-486a-bb7d-70a4e553e0cf-serving-cert\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215089 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/50003e5f-f76f-486a-bb7d-70a4e553e0cf-etcd-ca\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215106 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84edc937-d48f-439b-90a7-df4917ead6eb-config\") pod \"kube-controller-manager-operator-78b949d7b-5rhpw\" (UID: \"84edc937-d48f-439b-90a7-df4917ead6eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215123 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-encryption-config\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215141 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-config\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215158 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/fc341ad8-861e-427c-835c-b7bd7367b4a0-machine-approver-tls\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215174 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-audit\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215200 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-image-import-ca\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215218 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-config\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215233 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-audit-policies\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215250 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-encryption-config\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215265 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-etcd-serving-ca\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215282 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59f8296f-ab2b-40bc-af9c-80a1102321fa-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zzz2x\" (UID: \"59f8296f-ab2b-40bc-af9c-80a1102321fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215298 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50003e5f-f76f-486a-bb7d-70a4e553e0cf-etcd-client\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215323 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-default-certificate\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215340 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/975f5428-8720-4c72-8595-f213ef97888e-serving-cert\") pod \"openshift-config-operator-7777fb866f-c65h9\" (UID: \"975f5428-8720-4c72-8595-f213ef97888e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.215707 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.216114 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-node-pullsecrets\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.216215 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-njltp"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.216714 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-client-ca\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.216794 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/faa0cccb-7352-4752-8957-0d770dacaee5-audit-dir\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217473 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-audit-dir\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217517 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d9mkw\" (UID: \"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217555 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-config\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217582 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217605 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk2q6\" (UniqueName: \"kubernetes.io/projected/fc341ad8-861e-427c-835c-b7bd7367b4a0-kube-api-access-lk2q6\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217629 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217655 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217675 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jqjl\" (UniqueName: \"kubernetes.io/projected/a8054629-8c7a-4929-92c4-f720e3d30872-kube-api-access-2jqjl\") pod \"package-server-manager-789f6589d5-wdb6p\" (UID: \"a8054629-8c7a-4929-92c4-f720e3d30872\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217692 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217709 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnmpk\" (UniqueName: \"kubernetes.io/projected/ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8-kube-api-access-qnmpk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d9mkw\" (UID: \"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217726 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-config\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217745 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsvhn\" (UniqueName: \"kubernetes.io/projected/ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed-kube-api-access-fsvhn\") pod \"openshift-apiserver-operator-796bbdcf4f-zmv99\" (UID: \"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217745 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-trusted-ca\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217771 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqrpv\" (UniqueName: \"kubernetes.io/projected/59f8296f-ab2b-40bc-af9c-80a1102321fa-kube-api-access-bqrpv\") pod \"machine-config-controller-84d6567774-zzz2x\" (UID: \"59f8296f-ab2b-40bc-af9c-80a1102321fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217790 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217809 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxm72\" (UniqueName: \"kubernetes.io/projected/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-kube-api-access-wxm72\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217827 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-846fp\" (UniqueName: \"kubernetes.io/projected/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-kube-api-access-846fp\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217844 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/23523b1a-5f26-4f73-8ae9-a9735a83a7de-profile-collector-cert\") pod \"olm-operator-6b444d44fb-825zz\" (UID: \"23523b1a-5f26-4f73-8ae9-a9735a83a7de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217866 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-oauth-config\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217887 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-service-ca\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217914 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wptw\" (UniqueName: \"kubernetes.io/projected/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-kube-api-access-4wptw\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217932 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217931 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-config\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217950 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94036057-869a-49f0-84ca-8397bab8472c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xf9jw\" (UID: \"94036057-869a-49f0-84ca-8397bab8472c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217932 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/50003e5f-f76f-486a-bb7d-70a4e553e0cf-etcd-ca\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.217973 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zmv99\" (UID: \"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218018 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/975f5428-8720-4c72-8595-f213ef97888e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-c65h9\" (UID: \"975f5428-8720-4c72-8595-f213ef97888e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218095 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8054629-8c7a-4929-92c4-f720e3d30872-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wdb6p\" (UID: \"a8054629-8c7a-4929-92c4-f720e3d30872\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218127 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc0875cd-1663-45f8-a70b-96b0fb672d08-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6cgnd\" (UID: \"dc0875cd-1663-45f8-a70b-96b0fb672d08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218150 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc0875cd-1663-45f8-a70b-96b0fb672d08-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6cgnd\" (UID: \"dc0875cd-1663-45f8-a70b-96b0fb672d08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218180 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d00bf964-db4a-449d-ad97-d563b4b70f50-serving-cert\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218207 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-client-ca\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218233 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84edc937-d48f-439b-90a7-df4917ead6eb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5rhpw\" (UID: \"84edc937-d48f-439b-90a7-df4917ead6eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218275 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218292 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-config\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218486 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218530 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218666 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-images\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218682 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zmv99\" (UID: \"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218722 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/975f5428-8720-4c72-8595-f213ef97888e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-c65h9\" (UID: \"975f5428-8720-4c72-8595-f213ef97888e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.219191 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-audit-policies\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.219652 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-audit-dir\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.219639 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50003e5f-f76f-486a-bb7d-70a4e553e0cf-config\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.219842 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc341ad8-861e-427c-835c-b7bd7367b4a0-config\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.220121 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50003e5f-f76f-486a-bb7d-70a4e553e0cf-serving-cert\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.218300 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59f8296f-ab2b-40bc-af9c-80a1102321fa-proxy-tls\") pod \"machine-config-controller-84d6567774-zzz2x\" (UID: \"59f8296f-ab2b-40bc-af9c-80a1102321fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.220185 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-stats-auth\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.220216 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.220223 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-config\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.220240 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g9v6\" (UniqueName: \"kubernetes.io/projected/faa0cccb-7352-4752-8957-0d770dacaee5-kube-api-access-4g9v6\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.220385 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.220869 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.220945 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2839435a-a464-473c-94ad-7673ad00630b-serviceca\") pod \"image-pruner-29412000-5584k\" (UID: \"2839435a-a464-473c-94ad-7673ad00630b\") " pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.220966 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-client-ca\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.221253 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.221715 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-etcd-serving-ca\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.221739 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-audit\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.221824 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.222350 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-audit-policies\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.222520 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.222720 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-config\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.222968 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/50003e5f-f76f-486a-bb7d-70a4e553e0cf-etcd-service-ca\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.223176 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.224026 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-etcd-client\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.224226 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-audit-dir\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.224276 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.224654 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fc341ad8-861e-427c-835c-b7bd7367b4a0-auth-proxy-config\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.224665 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.224857 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-image-import-ca\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.224896 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.225288 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.225335 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59f8296f-ab2b-40bc-af9c-80a1102321fa-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zzz2x\" (UID: \"59f8296f-ab2b-40bc-af9c-80a1102321fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.226353 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-etcd-client\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.226489 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.226706 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d00bf964-db4a-449d-ad97-d563b4b70f50-serving-cert\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.227154 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.227157 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.227369 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50003e5f-f76f-486a-bb7d-70a4e553e0cf-etcd-client\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.227784 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-bbm58"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.228213 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.228322 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f33e67e7-af60-4c02-86f0-1bfd3de3d690-serving-cert\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.228527 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/975f5428-8720-4c72-8595-f213ef97888e-serving-cert\") pod \"openshift-config-operator-7777fb866f-c65h9\" (UID: \"975f5428-8720-4c72-8595-f213ef97888e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.229311 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-encryption-config\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.229518 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.230025 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vdwv6"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.230410 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59f8296f-ab2b-40bc-af9c-80a1102321fa-proxy-tls\") pod \"machine-config-controller-84d6567774-zzz2x\" (UID: \"59f8296f-ab2b-40bc-af9c-80a1102321fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.231080 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vdwv6" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.231713 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vldb5"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.232126 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-serving-cert\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.232158 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zmv99\" (UID: \"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.232172 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-serving-cert\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.232301 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/fc341ad8-861e-427c-835c-b7bd7367b4a0-machine-approver-tls\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.232585 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-metrics-tls\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.232955 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.233383 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vdwv6"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.233496 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.235005 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vldb5"] Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.235253 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.239552 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-encryption-config\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.266972 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.280314 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.295196 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.315114 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321128 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-proxy-tls\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321179 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94036057-869a-49f0-84ca-8397bab8472c-config\") pod \"kube-apiserver-operator-766d6c64bb-xf9jw\" (UID: \"94036057-869a-49f0-84ca-8397bab8472c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321209 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-metrics-certs\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321237 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77zl8\" (UniqueName: \"kubernetes.io/projected/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-kube-api-access-77zl8\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321273 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzc92\" (UniqueName: \"kubernetes.io/projected/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-kube-api-access-bzc92\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321300 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84edc937-d48f-439b-90a7-df4917ead6eb-config\") pod \"kube-controller-manager-operator-78b949d7b-5rhpw\" (UID: \"84edc937-d48f-439b-90a7-df4917ead6eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321342 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-default-certificate\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321370 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d9mkw\" (UID: \"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321421 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jqjl\" (UniqueName: \"kubernetes.io/projected/a8054629-8c7a-4929-92c4-f720e3d30872-kube-api-access-2jqjl\") pod \"package-server-manager-789f6589d5-wdb6p\" (UID: \"a8054629-8c7a-4929-92c4-f720e3d30872\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321462 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnmpk\" (UniqueName: \"kubernetes.io/projected/ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8-kube-api-access-qnmpk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d9mkw\" (UID: \"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321502 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-config\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321589 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/23523b1a-5f26-4f73-8ae9-a9735a83a7de-profile-collector-cert\") pod \"olm-operator-6b444d44fb-825zz\" (UID: \"23523b1a-5f26-4f73-8ae9-a9735a83a7de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321629 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-oauth-config\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321654 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-service-ca\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321691 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94036057-869a-49f0-84ca-8397bab8472c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xf9jw\" (UID: \"94036057-869a-49f0-84ca-8397bab8472c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321716 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc0875cd-1663-45f8-a70b-96b0fb672d08-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6cgnd\" (UID: \"dc0875cd-1663-45f8-a70b-96b0fb672d08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321745 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8054629-8c7a-4929-92c4-f720e3d30872-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wdb6p\" (UID: \"a8054629-8c7a-4929-92c4-f720e3d30872\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321774 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc0875cd-1663-45f8-a70b-96b0fb672d08-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6cgnd\" (UID: \"dc0875cd-1663-45f8-a70b-96b0fb672d08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321803 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84edc937-d48f-439b-90a7-df4917ead6eb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5rhpw\" (UID: \"84edc937-d48f-439b-90a7-df4917ead6eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321842 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-stats-auth\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321889 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc0875cd-1663-45f8-a70b-96b0fb672d08-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6cgnd\" (UID: \"dc0875cd-1663-45f8-a70b-96b0fb672d08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321929 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321954 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-trusted-ca-bundle\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.321977 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-service-ca-bundle\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322016 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94036057-869a-49f0-84ca-8397bab8472c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xf9jw\" (UID: \"94036057-869a-49f0-84ca-8397bab8472c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322039 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwh7l\" (UniqueName: \"kubernetes.io/projected/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-kube-api-access-fwh7l\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322068 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-oauth-serving-cert\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322103 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/23523b1a-5f26-4f73-8ae9-a9735a83a7de-srv-cert\") pod \"olm-operator-6b444d44fb-825zz\" (UID: \"23523b1a-5f26-4f73-8ae9-a9735a83a7de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322132 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4hvq\" (UniqueName: \"kubernetes.io/projected/23523b1a-5f26-4f73-8ae9-a9735a83a7de-kube-api-access-w4hvq\") pod \"olm-operator-6b444d44fb-825zz\" (UID: \"23523b1a-5f26-4f73-8ae9-a9735a83a7de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322171 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d9mkw\" (UID: \"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322198 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-serving-cert\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322226 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84edc937-d48f-439b-90a7-df4917ead6eb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5rhpw\" (UID: \"84edc937-d48f-439b-90a7-df4917ead6eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322255 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-images\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.322893 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94036057-869a-49f0-84ca-8397bab8472c-config\") pod \"kube-apiserver-operator-766d6c64bb-xf9jw\" (UID: \"94036057-869a-49f0-84ca-8397bab8472c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.323391 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-config\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.323644 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.326586 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc0875cd-1663-45f8-a70b-96b0fb672d08-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6cgnd\" (UID: \"dc0875cd-1663-45f8-a70b-96b0fb672d08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.329106 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-oauth-serving-cert\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.330004 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-service-ca\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.330149 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94036057-869a-49f0-84ca-8397bab8472c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-xf9jw\" (UID: \"94036057-869a-49f0-84ca-8397bab8472c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.330182 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-trusted-ca-bundle\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.331319 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc0875cd-1663-45f8-a70b-96b0fb672d08-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6cgnd\" (UID: \"dc0875cd-1663-45f8-a70b-96b0fb672d08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.331785 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-stats-auth\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.331952 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/23523b1a-5f26-4f73-8ae9-a9735a83a7de-srv-cert\") pod \"olm-operator-6b444d44fb-825zz\" (UID: \"23523b1a-5f26-4f73-8ae9-a9735a83a7de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.331971 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/23523b1a-5f26-4f73-8ae9-a9735a83a7de-profile-collector-cert\") pod \"olm-operator-6b444d44fb-825zz\" (UID: \"23523b1a-5f26-4f73-8ae9-a9735a83a7de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.332750 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-oauth-config\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.333336 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8054629-8c7a-4929-92c4-f720e3d30872-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wdb6p\" (UID: \"a8054629-8c7a-4929-92c4-f720e3d30872\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.335158 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.338632 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-serving-cert\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.355709 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.375400 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.386038 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-default-certificate\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.395104 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.406290 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-metrics-certs\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.415145 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.420193 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-service-ca-bundle\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.435020 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.455803 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.476516 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.495814 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.522070 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.535520 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.555565 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.575176 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.583194 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-images\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.595914 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.614975 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.625230 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-proxy-tls\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.655332 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.676328 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.695594 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.715399 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.735638 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.755809 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.762834 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84edc937-d48f-439b-90a7-df4917ead6eb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-5rhpw\" (UID: \"84edc937-d48f-439b-90a7-df4917ead6eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.775754 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.784179 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84edc937-d48f-439b-90a7-df4917ead6eb-config\") pod \"kube-controller-manager-operator-78b949d7b-5rhpw\" (UID: \"84edc937-d48f-439b-90a7-df4917ead6eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.795987 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.815365 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.835504 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.855323 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.875813 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.881725 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-d9mkw\" (UID: \"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.895907 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.915082 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.924152 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-d9mkw\" (UID: \"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.935536 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.955845 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 00:25:59 crc kubenswrapper[4912]: I1203 00:25:59.995731 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.016239 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.035162 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.055572 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.075387 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.094914 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.115298 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.133821 4912 request.go:700] Waited for 1.013649657s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/secrets?fieldSelector=metadata.name%3Ddns-operator-dockercfg-9mqw5&limit=500&resourceVersion=0 Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.135090 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.155973 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.174636 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.195383 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.214925 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.235476 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.254860 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.277390 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.295045 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.315167 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.335121 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.355197 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.374896 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.394843 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.422382 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.435163 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.436583 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.436740 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:26:00 crc kubenswrapper[4912]: E1203 00:26:00.436817 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:28:02.43678253 +0000 UTC m=+268.078803140 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.437140 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.442533 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.455655 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.475889 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.495527 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.515570 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.535597 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.538147 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.538232 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.541224 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.541632 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.560771 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.575362 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.595342 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.615041 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.635567 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.655240 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.675654 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.695239 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.715213 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.736052 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.755456 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.775772 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.795733 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.811033 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.815946 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.818167 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.835155 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.856101 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.894348 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt7ft\" (UniqueName: \"kubernetes.io/projected/2839435a-a464-473c-94ad-7673ad00630b-kube-api-access-zt7ft\") pod \"image-pruner-29412000-5584k\" (UID: \"2839435a-a464-473c-94ad-7673ad00630b\") " pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.933926 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgkwc\" (UniqueName: \"kubernetes.io/projected/50003e5f-f76f-486a-bb7d-70a4e553e0cf-kube-api-access-lgkwc\") pod \"etcd-operator-b45778765-7ckmb\" (UID: \"50003e5f-f76f-486a-bb7d-70a4e553e0cf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.946239 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.953958 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m76x\" (UniqueName: \"kubernetes.io/projected/61f377a2-06fc-427d-9c3e-38f9b3473e23-kube-api-access-7m76x\") pod \"downloads-7954f5f757-b5vr9\" (UID: \"61f377a2-06fc-427d-9c3e-38f9b3473e23\") " pod="openshift-console/downloads-7954f5f757-b5vr9" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.960411 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.969153 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g9n9\" (UniqueName: \"kubernetes.io/projected/d00bf964-db4a-449d-ad97-d563b4b70f50-kube-api-access-6g9n9\") pod \"route-controller-manager-6576b87f9c-8hfzk\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:26:00 crc kubenswrapper[4912]: I1203 00:26:00.987962 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g9v6\" (UniqueName: \"kubernetes.io/projected/faa0cccb-7352-4752-8957-0d770dacaee5-kube-api-access-4g9v6\") pod \"oauth-openshift-558db77b4-kd5n5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.008471 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j89rb\" (UniqueName: \"kubernetes.io/projected/ac6ae3d3-462a-4a82-a4a6-e225f12ce343-kube-api-access-j89rb\") pod \"machine-api-operator-5694c8668f-tqgnx\" (UID: \"ac6ae3d3-462a-4a82-a4a6-e225f12ce343\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.028818 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-846fp\" (UniqueName: \"kubernetes.io/projected/9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0-kube-api-access-846fp\") pod \"apiserver-7bbb656c7d-xlvsd\" (UID: \"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.030578 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.041212 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.053946 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsvhn\" (UniqueName: \"kubernetes.io/projected/ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed-kube-api-access-fsvhn\") pod \"openshift-apiserver-operator-796bbdcf4f-zmv99\" (UID: \"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.070090 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqrpv\" (UniqueName: \"kubernetes.io/projected/59f8296f-ab2b-40bc-af9c-80a1102321fa-kube-api-access-bqrpv\") pod \"machine-config-controller-84d6567774-zzz2x\" (UID: \"59f8296f-ab2b-40bc-af9c-80a1102321fa\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.089046 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.103475 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.108006 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxm72\" (UniqueName: \"kubernetes.io/projected/3938650f-17d8-4bb7-9fe7-5399eb3c4a5a-kube-api-access-wxm72\") pod \"ingress-operator-5b745b69d9-s2bl7\" (UID: \"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.121796 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.129078 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fffn6\" (UniqueName: \"kubernetes.io/projected/975f5428-8720-4c72-8595-f213ef97888e-kube-api-access-fffn6\") pod \"openshift-config-operator-7777fb866f-c65h9\" (UID: \"975f5428-8720-4c72-8595-f213ef97888e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.133917 4912 request.go:700] Waited for 1.909544691s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/serviceaccounts/machine-approver-sa/token Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.139214 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.150451 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk2q6\" (UniqueName: \"kubernetes.io/projected/fc341ad8-861e-427c-835c-b7bd7367b4a0-kube-api-access-lk2q6\") pod \"machine-approver-56656f9798-x4gmd\" (UID: \"fc341ad8-861e-427c-835c-b7bd7367b4a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.167845 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wptw\" (UniqueName: \"kubernetes.io/projected/3c4ce8f2-b6c1-4603-91b7-47e38cd693c2-kube-api-access-4wptw\") pod \"apiserver-76f77b778f-lg5fq\" (UID: \"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2\") " pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.174953 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.178756 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" Dec 03 00:26:01 crc kubenswrapper[4912]: W1203 00:26:01.191919 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc341ad8_861e_427c_835c_b7bd7367b4a0.slice/crio-7225cbd362231b51330d39e34793c0d6e2abb9de6b297cd277bf6efa6ec8c94c WatchSource:0}: Error finding container 7225cbd362231b51330d39e34793c0d6e2abb9de6b297cd277bf6efa6ec8c94c: Status 404 returned error can't find the container with id 7225cbd362231b51330d39e34793c0d6e2abb9de6b297cd277bf6efa6ec8c94c Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.196536 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.202088 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" event={"ID":"fc341ad8-861e-427c-835c-b7bd7367b4a0","Type":"ContainerStarted","Data":"7225cbd362231b51330d39e34793c0d6e2abb9de6b297cd277bf6efa6ec8c94c"} Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.207563 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b5vr9" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.215155 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.235374 4912 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.258848 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.261596 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.273113 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.275411 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.281134 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.308783 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnmpk\" (UniqueName: \"kubernetes.io/projected/ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8-kube-api-access-qnmpk\") pod \"openshift-controller-manager-operator-756b6f6bc6-d9mkw\" (UID: \"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.328507 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77zl8\" (UniqueName: \"kubernetes.io/projected/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-kube-api-access-77zl8\") pod \"console-f9d7485db-cg7ls\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.351230 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzc92\" (UniqueName: \"kubernetes.io/projected/bff567eb-a56f-42e6-9dc2-a58d8b2d8df0-kube-api-access-bzc92\") pod \"machine-config-operator-74547568cd-vtnw8\" (UID: \"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.363372 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.371146 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jqjl\" (UniqueName: \"kubernetes.io/projected/a8054629-8c7a-4929-92c4-f720e3d30872-kube-api-access-2jqjl\") pod \"package-server-manager-789f6589d5-wdb6p\" (UID: \"a8054629-8c7a-4929-92c4-f720e3d30872\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.379170 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.389085 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwh7l\" (UniqueName: \"kubernetes.io/projected/8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4-kube-api-access-fwh7l\") pod \"router-default-5444994796-dddxf\" (UID: \"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4\") " pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.398917 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.407316 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc0875cd-1663-45f8-a70b-96b0fb672d08-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6cgnd\" (UID: \"dc0875cd-1663-45f8-a70b-96b0fb672d08\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.428172 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4hvq\" (UniqueName: \"kubernetes.io/projected/23523b1a-5f26-4f73-8ae9-a9735a83a7de-kube-api-access-w4hvq\") pod \"olm-operator-6b444d44fb-825zz\" (UID: \"23523b1a-5f26-4f73-8ae9-a9735a83a7de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.430886 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.468293 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94036057-869a-49f0-84ca-8397bab8472c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-xf9jw\" (UID: \"94036057-869a-49f0-84ca-8397bab8472c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.895744 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.896186 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.897991 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/84edc937-d48f-439b-90a7-df4917ead6eb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-5rhpw\" (UID: \"84edc937-d48f-439b-90a7-df4917ead6eb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.898536 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.898587 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.899140 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qp8j\" (UniqueName: \"kubernetes.io/projected/f33e67e7-af60-4c02-86f0-1bfd3de3d690-kube-api-access-7qp8j\") pod \"controller-manager-879f6c89f-7x9ms\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.900117 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.904095 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.904696 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.905790 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-certificates\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.906671 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.907750 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-tls\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.907888 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a64deb49-4066-4d06-92fb-cd522ed7e496-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.909764 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:01 crc kubenswrapper[4912]: E1203 00:26:01.910363 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:02.410336845 +0000 UTC m=+148.052357405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.910563 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a64deb49-4066-4d06-92fb-cd522ed7e496-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.910631 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-bound-sa-token\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.910689 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-trusted-ca\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:01 crc kubenswrapper[4912]: I1203 00:26:01.910734 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmflw\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-kube-api-access-bmflw\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.011414 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.011547 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:02.511528251 +0000 UTC m=+148.153548811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.011854 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28f13442-6621-41b8-9d00-b6607702fa48-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bq8vd\" (UID: \"28f13442-6621-41b8-9d00-b6607702fa48\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.011896 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-trusted-ca\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.011920 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8b45be-2bdf-447c-82f7-758871f840e3-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012177 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbhjh\" (UniqueName: \"kubernetes.io/projected/a1e2de7d-cdd8-433b-b42c-57e6cb6d8679-kube-api-access-hbhjh\") pod \"service-ca-9c57cc56f-mv4s4\" (UID: \"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679\") " pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012206 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8b45be-2bdf-447c-82f7-758871f840e3-serving-cert\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012275 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50d1bf1c-014f-4900-a835-96cf690299e9-srv-cert\") pod \"catalog-operator-68c6474976-gm5mt\" (UID: \"50d1bf1c-014f-4900-a835-96cf690299e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012347 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0f44254a-d149-4a4b-aec0-1a2613d47f81-metrics-tls\") pod \"dns-operator-744455d44c-njltp\" (UID: \"0f44254a-d149-4a4b-aec0-1a2613d47f81\") " pod="openshift-dns-operator/dns-operator-744455d44c-njltp" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012368 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stj2t\" (UniqueName: \"kubernetes.io/projected/8f8b45be-2bdf-447c-82f7-758871f840e3-kube-api-access-stj2t\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012593 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f9e9c612-02e2-4817-920f-e500720bd6ff-node-bootstrap-token\") pod \"machine-config-server-n8lb6\" (UID: \"f9e9c612-02e2-4817-920f-e500720bd6ff\") " pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012611 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f9e9c612-02e2-4817-920f-e500720bd6ff-certs\") pod \"machine-config-server-n8lb6\" (UID: \"f9e9c612-02e2-4817-920f-e500720bd6ff\") " pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012628 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a64deb49-4066-4d06-92fb-cd522ed7e496-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012687 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-trusted-ca\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012887 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpn5j\" (UniqueName: \"kubernetes.io/projected/896d21a9-3b77-4571-a4c2-edc02fdf6db2-kube-api-access-hpn5j\") pod \"kube-storage-version-migrator-operator-b67b599dd-5fg2k\" (UID: \"896d21a9-3b77-4571-a4c2-edc02fdf6db2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012914 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd032b60-199f-4f51-b635-11666dc68b7c-metrics-tls\") pod \"dns-default-vdwv6\" (UID: \"fd032b60-199f-4f51-b635-11666dc68b7c\") " pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012932 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-mountpoint-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.012988 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dgskz\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013007 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8b45be-2bdf-447c-82f7-758871f840e3-service-ca-bundle\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013022 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/edb6b379-aeff-470a-a09d-540d6acc0efe-tmpfs\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013064 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a1e2de7d-cdd8-433b-b42c-57e6cb6d8679-signing-cabundle\") pod \"service-ca-9c57cc56f-mv4s4\" (UID: \"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679\") " pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013138 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5r4f\" (UniqueName: \"kubernetes.io/projected/fd032b60-199f-4f51-b635-11666dc68b7c-kube-api-access-b5r4f\") pod \"dns-default-vdwv6\" (UID: \"fd032b60-199f-4f51-b635-11666dc68b7c\") " pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013184 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt8bh\" (UniqueName: \"kubernetes.io/projected/82f9b623-024f-45e6-b876-f9718bc8aba6-kube-api-access-bt8bh\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013271 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-serving-cert\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013310 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrx92\" (UniqueName: \"kubernetes.io/projected/cdbf7d39-1149-49d5-956b-0d2d31362c0e-kube-api-access-qrx92\") pod \"migrator-59844c95c7-8cc6v\" (UID: \"cdbf7d39-1149-49d5-956b-0d2d31362c0e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013377 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bd6c8225-9d04-4ecf-9163-21508a4fd09a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013408 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38ad0fca-841e-4f4f-a6b3-ad172aa5b781-serving-cert\") pod \"service-ca-operator-777779d784-bbm58\" (UID: \"38ad0fca-841e-4f4f-a6b3-ad172aa5b781\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013447 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjbml\" (UniqueName: \"kubernetes.io/projected/f6a647db-07ba-49f6-9150-0a038d4fbee8-kube-api-access-mjbml\") pod \"collect-profiles-29412015-66lp4\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013486 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn7pk\" (UniqueName: \"kubernetes.io/projected/0f44254a-d149-4a4b-aec0-1a2613d47f81-kube-api-access-dn7pk\") pod \"dns-operator-744455d44c-njltp\" (UID: \"0f44254a-d149-4a4b-aec0-1a2613d47f81\") " pod="openshift-dns-operator/dns-operator-744455d44c-njltp" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013550 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-tls\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013573 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-config\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013651 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnwsw\" (UniqueName: \"kubernetes.io/projected/edb6b379-aeff-470a-a09d-540d6acc0efe-kube-api-access-rnwsw\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013678 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9-cert\") pod \"ingress-canary-gqt86\" (UID: \"9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9\") " pod="openshift-ingress-canary/ingress-canary-gqt86" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013706 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a64deb49-4066-4d06-92fb-cd522ed7e496-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013754 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcqx5\" (UniqueName: \"kubernetes.io/projected/c916e93d-66d8-436b-841e-c7a92f823912-kube-api-access-hcqx5\") pod \"marketplace-operator-79b997595-dgskz\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013778 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6a647db-07ba-49f6-9150-0a038d4fbee8-secret-volume\") pod \"collect-profiles-29412015-66lp4\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013808 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn2g8\" (UniqueName: \"kubernetes.io/projected/9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9-kube-api-access-qn2g8\") pod \"ingress-canary-gqt86\" (UID: \"9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9\") " pod="openshift-ingress-canary/ingress-canary-gqt86" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.013905 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8990726-b55a-4424-8d40-18132fbc955c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t69lf\" (UID: \"e8990726-b55a-4424-8d40-18132fbc955c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.014247 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.014297 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kzdp\" (UniqueName: \"kubernetes.io/projected/28f13442-6621-41b8-9d00-b6607702fa48-kube-api-access-2kzdp\") pod \"cluster-samples-operator-665b6dd947-bq8vd\" (UID: \"28f13442-6621-41b8-9d00-b6607702fa48\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.014320 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-csi-data-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.014369 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a64deb49-4066-4d06-92fb-cd522ed7e496-ca-trust-extracted\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.014559 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-registration-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.014755 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:02.514738504 +0000 UTC m=+148.156759054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.014817 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9s9b\" (UniqueName: \"kubernetes.io/projected/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-kube-api-access-d9s9b\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.014899 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50d1bf1c-014f-4900-a835-96cf690299e9-profile-collector-cert\") pod \"catalog-operator-68c6474976-gm5mt\" (UID: \"50d1bf1c-014f-4900-a835-96cf690299e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.014953 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-bound-sa-token\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.014989 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/896d21a9-3b77-4571-a4c2-edc02fdf6db2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5fg2k\" (UID: \"896d21a9-3b77-4571-a4c2-edc02fdf6db2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015014 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-socket-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015118 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-trusted-ca\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015165 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmflw\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-kube-api-access-bmflw\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015196 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2fw6\" (UniqueName: \"kubernetes.io/projected/38ad0fca-841e-4f4f-a6b3-ad172aa5b781-kube-api-access-l2fw6\") pod \"service-ca-operator-777779d784-bbm58\" (UID: \"38ad0fca-841e-4f4f-a6b3-ad172aa5b781\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015254 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/edb6b379-aeff-470a-a09d-540d6acc0efe-apiservice-cert\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015325 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/896d21a9-3b77-4571-a4c2-edc02fdf6db2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5fg2k\" (UID: \"896d21a9-3b77-4571-a4c2-edc02fdf6db2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015493 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v49ml\" (UniqueName: \"kubernetes.io/projected/cff6f92d-904b-43ba-a0d7-2bae917cdcd8-kube-api-access-v49ml\") pod \"control-plane-machine-set-operator-78cbb6b69f-g5h6h\" (UID: \"cff6f92d-904b-43ba-a0d7-2bae917cdcd8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015596 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr9gb\" (UniqueName: \"kubernetes.io/projected/bd6c8225-9d04-4ecf-9163-21508a4fd09a-kube-api-access-xr9gb\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015729 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-certificates\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015775 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd6c8225-9d04-4ecf-9163-21508a4fd09a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.015841 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr9hv\" (UniqueName: \"kubernetes.io/projected/f9e9c612-02e2-4817-920f-e500720bd6ff-kube-api-access-jr9hv\") pod \"machine-config-server-n8lb6\" (UID: \"f9e9c612-02e2-4817-920f-e500720bd6ff\") " pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016220 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6a647db-07ba-49f6-9150-0a038d4fbee8-config-volume\") pod \"collect-profiles-29412015-66lp4\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016258 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cff6f92d-904b-43ba-a0d7-2bae917cdcd8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-g5h6h\" (UID: \"cff6f92d-904b-43ba-a0d7-2bae917cdcd8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016347 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd6c8225-9d04-4ecf-9163-21508a4fd09a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016381 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/edb6b379-aeff-470a-a09d-540d6acc0efe-webhook-cert\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016603 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jpvt\" (UniqueName: \"kubernetes.io/projected/50d1bf1c-014f-4900-a835-96cf690299e9-kube-api-access-8jpvt\") pod \"catalog-operator-68c6474976-gm5mt\" (UID: \"50d1bf1c-014f-4900-a835-96cf690299e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016674 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dgskz\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016692 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f8b45be-2bdf-447c-82f7-758871f840e3-config\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016706 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-plugins-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016726 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt2js\" (UniqueName: \"kubernetes.io/projected/e8990726-b55a-4424-8d40-18132fbc955c-kube-api-access-wt2js\") pod \"multus-admission-controller-857f4d67dd-t69lf\" (UID: \"e8990726-b55a-4424-8d40-18132fbc955c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.016969 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd032b60-199f-4f51-b635-11666dc68b7c-config-volume\") pod \"dns-default-vdwv6\" (UID: \"fd032b60-199f-4f51-b635-11666dc68b7c\") " pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.017006 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ad0fca-841e-4f4f-a6b3-ad172aa5b781-config\") pod \"service-ca-operator-777779d784-bbm58\" (UID: \"38ad0fca-841e-4f4f-a6b3-ad172aa5b781\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.017034 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a1e2de7d-cdd8-433b-b42c-57e6cb6d8679-signing-key\") pod \"service-ca-9c57cc56f-mv4s4\" (UID: \"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679\") " pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.017673 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-certificates\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.018393 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a64deb49-4066-4d06-92fb-cd522ed7e496-installation-pull-secrets\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.019247 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-tls\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.050825 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.098317 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-bound-sa-token\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.107929 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmflw\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-kube-api-access-bmflw\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119059 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.119534 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:02.619511573 +0000 UTC m=+148.261532143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119618 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/edb6b379-aeff-470a-a09d-540d6acc0efe-webhook-cert\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119651 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd6c8225-9d04-4ecf-9163-21508a4fd09a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119678 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jpvt\" (UniqueName: \"kubernetes.io/projected/50d1bf1c-014f-4900-a835-96cf690299e9-kube-api-access-8jpvt\") pod \"catalog-operator-68c6474976-gm5mt\" (UID: \"50d1bf1c-014f-4900-a835-96cf690299e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119725 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dgskz\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119745 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f8b45be-2bdf-447c-82f7-758871f840e3-config\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119766 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-plugins-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119791 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt2js\" (UniqueName: \"kubernetes.io/projected/e8990726-b55a-4424-8d40-18132fbc955c-kube-api-access-wt2js\") pod \"multus-admission-controller-857f4d67dd-t69lf\" (UID: \"e8990726-b55a-4424-8d40-18132fbc955c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119816 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd032b60-199f-4f51-b635-11666dc68b7c-config-volume\") pod \"dns-default-vdwv6\" (UID: \"fd032b60-199f-4f51-b635-11666dc68b7c\") " pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119844 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ad0fca-841e-4f4f-a6b3-ad172aa5b781-config\") pod \"service-ca-operator-777779d784-bbm58\" (UID: \"38ad0fca-841e-4f4f-a6b3-ad172aa5b781\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119867 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a1e2de7d-cdd8-433b-b42c-57e6cb6d8679-signing-key\") pod \"service-ca-9c57cc56f-mv4s4\" (UID: \"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679\") " pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119891 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28f13442-6621-41b8-9d00-b6607702fa48-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bq8vd\" (UID: \"28f13442-6621-41b8-9d00-b6607702fa48\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119914 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-trusted-ca\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119942 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8b45be-2bdf-447c-82f7-758871f840e3-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.119997 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbhjh\" (UniqueName: \"kubernetes.io/projected/a1e2de7d-cdd8-433b-b42c-57e6cb6d8679-kube-api-access-hbhjh\") pod \"service-ca-9c57cc56f-mv4s4\" (UID: \"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679\") " pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.120033 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8b45be-2bdf-447c-82f7-758871f840e3-serving-cert\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.120056 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50d1bf1c-014f-4900-a835-96cf690299e9-srv-cert\") pod \"catalog-operator-68c6474976-gm5mt\" (UID: \"50d1bf1c-014f-4900-a835-96cf690299e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.120084 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0f44254a-d149-4a4b-aec0-1a2613d47f81-metrics-tls\") pod \"dns-operator-744455d44c-njltp\" (UID: \"0f44254a-d149-4a4b-aec0-1a2613d47f81\") " pod="openshift-dns-operator/dns-operator-744455d44c-njltp" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.122319 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/38ad0fca-841e-4f4f-a6b3-ad172aa5b781-config\") pod \"service-ca-operator-777779d784-bbm58\" (UID: \"38ad0fca-841e-4f4f-a6b3-ad172aa5b781\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.120107 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stj2t\" (UniqueName: \"kubernetes.io/projected/8f8b45be-2bdf-447c-82f7-758871f840e3-kube-api-access-stj2t\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123069 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f9e9c612-02e2-4817-920f-e500720bd6ff-node-bootstrap-token\") pod \"machine-config-server-n8lb6\" (UID: \"f9e9c612-02e2-4817-920f-e500720bd6ff\") " pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123091 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f9e9c612-02e2-4817-920f-e500720bd6ff-certs\") pod \"machine-config-server-n8lb6\" (UID: \"f9e9c612-02e2-4817-920f-e500720bd6ff\") " pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123108 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f8b45be-2bdf-447c-82f7-758871f840e3-config\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123129 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpn5j\" (UniqueName: \"kubernetes.io/projected/896d21a9-3b77-4571-a4c2-edc02fdf6db2-kube-api-access-hpn5j\") pod \"kube-storage-version-migrator-operator-b67b599dd-5fg2k\" (UID: \"896d21a9-3b77-4571-a4c2-edc02fdf6db2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123168 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd032b60-199f-4f51-b635-11666dc68b7c-metrics-tls\") pod \"dns-default-vdwv6\" (UID: \"fd032b60-199f-4f51-b635-11666dc68b7c\") " pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123188 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-mountpoint-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123212 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dgskz\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123234 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8b45be-2bdf-447c-82f7-758871f840e3-service-ca-bundle\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123269 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/edb6b379-aeff-470a-a09d-540d6acc0efe-tmpfs\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123290 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a1e2de7d-cdd8-433b-b42c-57e6cb6d8679-signing-cabundle\") pod \"service-ca-9c57cc56f-mv4s4\" (UID: \"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679\") " pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123319 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5r4f\" (UniqueName: \"kubernetes.io/projected/fd032b60-199f-4f51-b635-11666dc68b7c-kube-api-access-b5r4f\") pod \"dns-default-vdwv6\" (UID: \"fd032b60-199f-4f51-b635-11666dc68b7c\") " pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123339 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt8bh\" (UniqueName: \"kubernetes.io/projected/82f9b623-024f-45e6-b876-f9718bc8aba6-kube-api-access-bt8bh\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123379 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-serving-cert\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123404 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrx92\" (UniqueName: \"kubernetes.io/projected/cdbf7d39-1149-49d5-956b-0d2d31362c0e-kube-api-access-qrx92\") pod \"migrator-59844c95c7-8cc6v\" (UID: \"cdbf7d39-1149-49d5-956b-0d2d31362c0e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123448 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bd6c8225-9d04-4ecf-9163-21508a4fd09a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123473 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38ad0fca-841e-4f4f-a6b3-ad172aa5b781-serving-cert\") pod \"service-ca-operator-777779d784-bbm58\" (UID: \"38ad0fca-841e-4f4f-a6b3-ad172aa5b781\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123503 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjbml\" (UniqueName: \"kubernetes.io/projected/f6a647db-07ba-49f6-9150-0a038d4fbee8-kube-api-access-mjbml\") pod \"collect-profiles-29412015-66lp4\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123534 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn7pk\" (UniqueName: \"kubernetes.io/projected/0f44254a-d149-4a4b-aec0-1a2613d47f81-kube-api-access-dn7pk\") pod \"dns-operator-744455d44c-njltp\" (UID: \"0f44254a-d149-4a4b-aec0-1a2613d47f81\") " pod="openshift-dns-operator/dns-operator-744455d44c-njltp" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123571 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-config\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123595 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnwsw\" (UniqueName: \"kubernetes.io/projected/edb6b379-aeff-470a-a09d-540d6acc0efe-kube-api-access-rnwsw\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123619 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9-cert\") pod \"ingress-canary-gqt86\" (UID: \"9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9\") " pod="openshift-ingress-canary/ingress-canary-gqt86" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123689 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcqx5\" (UniqueName: \"kubernetes.io/projected/c916e93d-66d8-436b-841e-c7a92f823912-kube-api-access-hcqx5\") pod \"marketplace-operator-79b997595-dgskz\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123708 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6a647db-07ba-49f6-9150-0a038d4fbee8-secret-volume\") pod \"collect-profiles-29412015-66lp4\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123728 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn2g8\" (UniqueName: \"kubernetes.io/projected/9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9-kube-api-access-qn2g8\") pod \"ingress-canary-gqt86\" (UID: \"9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9\") " pod="openshift-ingress-canary/ingress-canary-gqt86" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123755 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8990726-b55a-4424-8d40-18132fbc955c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t69lf\" (UID: \"e8990726-b55a-4424-8d40-18132fbc955c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123775 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-csi-data-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123801 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123819 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kzdp\" (UniqueName: \"kubernetes.io/projected/28f13442-6621-41b8-9d00-b6607702fa48-kube-api-access-2kzdp\") pod \"cluster-samples-operator-665b6dd947-bq8vd\" (UID: \"28f13442-6621-41b8-9d00-b6607702fa48\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123839 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-registration-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123868 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9s9b\" (UniqueName: \"kubernetes.io/projected/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-kube-api-access-d9s9b\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123895 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50d1bf1c-014f-4900-a835-96cf690299e9-profile-collector-cert\") pod \"catalog-operator-68c6474976-gm5mt\" (UID: \"50d1bf1c-014f-4900-a835-96cf690299e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123923 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/896d21a9-3b77-4571-a4c2-edc02fdf6db2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5fg2k\" (UID: \"896d21a9-3b77-4571-a4c2-edc02fdf6db2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.123946 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-socket-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.124000 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2fw6\" (UniqueName: \"kubernetes.io/projected/38ad0fca-841e-4f4f-a6b3-ad172aa5b781-kube-api-access-l2fw6\") pod \"service-ca-operator-777779d784-bbm58\" (UID: \"38ad0fca-841e-4f4f-a6b3-ad172aa5b781\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.124024 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/edb6b379-aeff-470a-a09d-540d6acc0efe-apiservice-cert\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.124048 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/896d21a9-3b77-4571-a4c2-edc02fdf6db2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5fg2k\" (UID: \"896d21a9-3b77-4571-a4c2-edc02fdf6db2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.124073 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v49ml\" (UniqueName: \"kubernetes.io/projected/cff6f92d-904b-43ba-a0d7-2bae917cdcd8-kube-api-access-v49ml\") pod \"control-plane-machine-set-operator-78cbb6b69f-g5h6h\" (UID: \"cff6f92d-904b-43ba-a0d7-2bae917cdcd8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.124100 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr9gb\" (UniqueName: \"kubernetes.io/projected/bd6c8225-9d04-4ecf-9163-21508a4fd09a-kube-api-access-xr9gb\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.124124 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd6c8225-9d04-4ecf-9163-21508a4fd09a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.124148 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr9hv\" (UniqueName: \"kubernetes.io/projected/f9e9c612-02e2-4817-920f-e500720bd6ff-kube-api-access-jr9hv\") pod \"machine-config-server-n8lb6\" (UID: \"f9e9c612-02e2-4817-920f-e500720bd6ff\") " pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.124177 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6a647db-07ba-49f6-9150-0a038d4fbee8-config-volume\") pod \"collect-profiles-29412015-66lp4\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.124199 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cff6f92d-904b-43ba-a0d7-2bae917cdcd8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-g5h6h\" (UID: \"cff6f92d-904b-43ba-a0d7-2bae917cdcd8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.126391 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/edb6b379-aeff-470a-a09d-540d6acc0efe-webhook-cert\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.127314 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-plugins-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.128070 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dgskz\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.129299 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-config\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.130758 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-socket-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.131018 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd032b60-199f-4f51-b635-11666dc68b7c-config-volume\") pod \"dns-default-vdwv6\" (UID: \"fd032b60-199f-4f51-b635-11666dc68b7c\") " pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.131360 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8b45be-2bdf-447c-82f7-758871f840e3-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.131560 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a1e2de7d-cdd8-433b-b42c-57e6cb6d8679-signing-cabundle\") pod \"service-ca-9c57cc56f-mv4s4\" (UID: \"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679\") " pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.131677 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dgskz\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.132175 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8f8b45be-2bdf-447c-82f7-758871f840e3-service-ca-bundle\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.132566 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/edb6b379-aeff-470a-a09d-540d6acc0efe-tmpfs\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.132912 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-registration-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.133208 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/28f13442-6621-41b8-9d00-b6607702fa48-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bq8vd\" (UID: \"28f13442-6621-41b8-9d00-b6607702fa48\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.133671 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9-cert\") pod \"ingress-canary-gqt86\" (UID: \"9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9\") " pod="openshift-ingress-canary/ingress-canary-gqt86" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.134685 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd6c8225-9d04-4ecf-9163-21508a4fd09a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.138139 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-mountpoint-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.138231 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/82f9b623-024f-45e6-b876-f9718bc8aba6-csi-data-dir\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.138866 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6a647db-07ba-49f6-9150-0a038d4fbee8-config-volume\") pod \"collect-profiles-29412015-66lp4\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.141291 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/896d21a9-3b77-4571-a4c2-edc02fdf6db2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-5fg2k\" (UID: \"896d21a9-3b77-4571-a4c2-edc02fdf6db2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.141649 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-trusted-ca\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.144216 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:02.644171965 +0000 UTC m=+148.286192525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.147402 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a1e2de7d-cdd8-433b-b42c-57e6cb6d8679-signing-key\") pod \"service-ca-9c57cc56f-mv4s4\" (UID: \"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679\") " pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.147838 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cff6f92d-904b-43ba-a0d7-2bae917cdcd8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-g5h6h\" (UID: \"cff6f92d-904b-43ba-a0d7-2bae917cdcd8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.149370 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/edb6b379-aeff-470a-a09d-540d6acc0efe-apiservice-cert\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.153083 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-serving-cert\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.156153 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f8b45be-2bdf-447c-82f7-758871f840e3-serving-cert\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.162415 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/896d21a9-3b77-4571-a4c2-edc02fdf6db2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-5fg2k\" (UID: \"896d21a9-3b77-4571-a4c2-edc02fdf6db2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.168365 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/50d1bf1c-014f-4900-a835-96cf690299e9-srv-cert\") pod \"catalog-operator-68c6474976-gm5mt\" (UID: \"50d1bf1c-014f-4900-a835-96cf690299e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.168700 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/38ad0fca-841e-4f4f-a6b3-ad172aa5b781-serving-cert\") pod \"service-ca-operator-777779d784-bbm58\" (UID: \"38ad0fca-841e-4f4f-a6b3-ad172aa5b781\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.169821 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6a647db-07ba-49f6-9150-0a038d4fbee8-secret-volume\") pod \"collect-profiles-29412015-66lp4\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.173281 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd6c8225-9d04-4ecf-9163-21508a4fd09a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.179837 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fd032b60-199f-4f51-b635-11666dc68b7c-metrics-tls\") pod \"dns-default-vdwv6\" (UID: \"fd032b60-199f-4f51-b635-11666dc68b7c\") " pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.181170 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bd6c8225-9d04-4ecf-9163-21508a4fd09a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.182693 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0f44254a-d149-4a4b-aec0-1a2613d47f81-metrics-tls\") pod \"dns-operator-744455d44c-njltp\" (UID: \"0f44254a-d149-4a4b-aec0-1a2613d47f81\") " pod="openshift-dns-operator/dns-operator-744455d44c-njltp" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.183589 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e8990726-b55a-4424-8d40-18132fbc955c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-t69lf\" (UID: \"e8990726-b55a-4424-8d40-18132fbc955c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.189595 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f9e9c612-02e2-4817-920f-e500720bd6ff-node-bootstrap-token\") pod \"machine-config-server-n8lb6\" (UID: \"f9e9c612-02e2-4817-920f-e500720bd6ff\") " pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.189639 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/50d1bf1c-014f-4900-a835-96cf690299e9-profile-collector-cert\") pod \"catalog-operator-68c6474976-gm5mt\" (UID: \"50d1bf1c-014f-4900-a835-96cf690299e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.191259 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f9e9c612-02e2-4817-920f-e500720bd6ff-certs\") pod \"machine-config-server-n8lb6\" (UID: \"f9e9c612-02e2-4817-920f-e500720bd6ff\") " pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.226103 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.226648 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:02.72662825 +0000 UTC m=+148.368648810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.240213 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jpvt\" (UniqueName: \"kubernetes.io/projected/50d1bf1c-014f-4900-a835-96cf690299e9-kube-api-access-8jpvt\") pod \"catalog-operator-68c6474976-gm5mt\" (UID: \"50d1bf1c-014f-4900-a835-96cf690299e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.255284 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9s9b\" (UniqueName: \"kubernetes.io/projected/54c92a35-71f5-4fe9-9936-c7a3ea0990c8-kube-api-access-d9s9b\") pod \"console-operator-58897d9998-4pwws\" (UID: \"54c92a35-71f5-4fe9-9936-c7a3ea0990c8\") " pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.292302 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.293233 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnwsw\" (UniqueName: \"kubernetes.io/projected/edb6b379-aeff-470a-a09d-540d6acc0efe-kube-api-access-rnwsw\") pod \"packageserver-d55dfcdfc-cns6n\" (UID: \"edb6b379-aeff-470a-a09d-540d6acc0efe\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.303923 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrx92\" (UniqueName: \"kubernetes.io/projected/cdbf7d39-1149-49d5-956b-0d2d31362c0e-kube-api-access-qrx92\") pod \"migrator-59844c95c7-8cc6v\" (UID: \"cdbf7d39-1149-49d5-956b-0d2d31362c0e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.305752 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt2js\" (UniqueName: \"kubernetes.io/projected/e8990726-b55a-4424-8d40-18132fbc955c-kube-api-access-wt2js\") pod \"multus-admission-controller-857f4d67dd-t69lf\" (UID: \"e8990726-b55a-4424-8d40-18132fbc955c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.315216 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbhjh\" (UniqueName: \"kubernetes.io/projected/a1e2de7d-cdd8-433b-b42c-57e6cb6d8679-kube-api-access-hbhjh\") pod \"service-ca-9c57cc56f-mv4s4\" (UID: \"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679\") " pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.327952 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.328668 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:02.82865586 +0000 UTC m=+148.470676420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.349100 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v49ml\" (UniqueName: \"kubernetes.io/projected/cff6f92d-904b-43ba-a0d7-2bae917cdcd8-kube-api-access-v49ml\") pod \"control-plane-machine-set-operator-78cbb6b69f-g5h6h\" (UID: \"cff6f92d-904b-43ba-a0d7-2bae917cdcd8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.361085 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2fw6\" (UniqueName: \"kubernetes.io/projected/38ad0fca-841e-4f4f-a6b3-ad172aa5b781-kube-api-access-l2fw6\") pod \"service-ca-operator-777779d784-bbm58\" (UID: \"38ad0fca-841e-4f4f-a6b3-ad172aa5b781\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.367360 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.374527 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.394147 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcqx5\" (UniqueName: \"kubernetes.io/projected/c916e93d-66d8-436b-841e-c7a92f823912-kube-api-access-hcqx5\") pod \"marketplace-operator-79b997595-dgskz\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.397181 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.401811 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjbml\" (UniqueName: \"kubernetes.io/projected/f6a647db-07ba-49f6-9150-0a038d4fbee8-kube-api-access-mjbml\") pod \"collect-profiles-29412015-66lp4\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.404153 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr9gb\" (UniqueName: \"kubernetes.io/projected/bd6c8225-9d04-4ecf-9163-21508a4fd09a-kube-api-access-xr9gb\") pod \"cluster-image-registry-operator-dc59b4c8b-xf4wl\" (UID: \"bd6c8225-9d04-4ecf-9163-21508a4fd09a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.404367 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.410367 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.410750 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn7pk\" (UniqueName: \"kubernetes.io/projected/0f44254a-d149-4a4b-aec0-1a2613d47f81-kube-api-access-dn7pk\") pod \"dns-operator-744455d44c-njltp\" (UID: \"0f44254a-d149-4a4b-aec0-1a2613d47f81\") " pod="openshift-dns-operator/dns-operator-744455d44c-njltp" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.417510 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-njltp" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.424127 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.429743 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.430083 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:02.930067453 +0000 UTC m=+148.572088013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.431520 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kzdp\" (UniqueName: \"kubernetes.io/projected/28f13442-6621-41b8-9d00-b6607702fa48-kube-api-access-2kzdp\") pod \"cluster-samples-operator-665b6dd947-bq8vd\" (UID: \"28f13442-6621-41b8-9d00-b6607702fa48\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.433856 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.441227 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.453657 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpn5j\" (UniqueName: \"kubernetes.io/projected/896d21a9-3b77-4571-a4c2-edc02fdf6db2-kube-api-access-hpn5j\") pod \"kube-storage-version-migrator-operator-b67b599dd-5fg2k\" (UID: \"896d21a9-3b77-4571-a4c2-edc02fdf6db2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.477649 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stj2t\" (UniqueName: \"kubernetes.io/projected/8f8b45be-2bdf-447c-82f7-758871f840e3-kube-api-access-stj2t\") pod \"authentication-operator-69f744f599-4z6ww\" (UID: \"8f8b45be-2bdf-447c-82f7-758871f840e3\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.496995 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.497740 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn2g8\" (UniqueName: \"kubernetes.io/projected/9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9-kube-api-access-qn2g8\") pod \"ingress-canary-gqt86\" (UID: \"9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9\") " pod="openshift-ingress-canary/ingress-canary-gqt86" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.508954 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.542508 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.542987 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.042973796 +0000 UTC m=+148.684994356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.550347 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr9hv\" (UniqueName: \"kubernetes.io/projected/f9e9c612-02e2-4817-920f-e500720bd6ff-kube-api-access-jr9hv\") pod \"machine-config-server-n8lb6\" (UID: \"f9e9c612-02e2-4817-920f-e500720bd6ff\") " pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.554162 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5r4f\" (UniqueName: \"kubernetes.io/projected/fd032b60-199f-4f51-b635-11666dc68b7c-kube-api-access-b5r4f\") pod \"dns-default-vdwv6\" (UID: \"fd032b60-199f-4f51-b635-11666dc68b7c\") " pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.561495 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt8bh\" (UniqueName: \"kubernetes.io/projected/82f9b623-024f-45e6-b876-f9718bc8aba6-kube-api-access-bt8bh\") pod \"csi-hostpathplugin-vldb5\" (UID: \"82f9b623-024f-45e6-b876-f9718bc8aba6\") " pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.630631 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.645913 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.646321 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.146303204 +0000 UTC m=+148.788323774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.664929 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.683204 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.750619 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-gqt86" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.753693 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.758261 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.258239011 +0000 UTC m=+148.900259571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.759170 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-n8lb6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.770557 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.801162 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vldb5" Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.858376 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.858698 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.358674825 +0000 UTC m=+149.000695385 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.858905 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.859198 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.359186239 +0000 UTC m=+149.001206799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: W1203 00:26:02.937448 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9e9c612_02e2_4817_920f_e500720bd6ff.slice/crio-07b3ceeb2a1208b5e34e38c2be9f81df6d040deebdeccc979be2aa7f3fdb9662 WatchSource:0}: Error finding container 07b3ceeb2a1208b5e34e38c2be9f81df6d040deebdeccc979be2aa7f3fdb9662: Status 404 returned error can't find the container with id 07b3ceeb2a1208b5e34e38c2be9f81df6d040deebdeccc979be2aa7f3fdb9662 Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.960997 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.961176 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.461145387 +0000 UTC m=+149.103165957 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:02 crc kubenswrapper[4912]: I1203 00:26:02.961249 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:02 crc kubenswrapper[4912]: E1203 00:26:02.961599 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.46158623 +0000 UTC m=+149.103606790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.061747 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.062115 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.562101116 +0000 UTC m=+149.204121676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.162993 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.163311 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.663297722 +0000 UTC m=+149.305318282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.250222 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"066a1d616a6288f142a36a9fa865bb652a68563ddd58430641cfac8c70f60bd2"} Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.250528 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"dee010b9553c013eb19ed50c567557bea2c602365450be953dba392f144ccaed"} Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.251467 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-dddxf" event={"ID":"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4","Type":"ContainerStarted","Data":"e2e9446bc39c457db1245548172437f25d88d072e8b8b7c92ad84aa084a33560"} Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.251508 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-dddxf" event={"ID":"8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4","Type":"ContainerStarted","Data":"864ca6b736b6964f74afa7f2d01a4e42726144ea3389f2864b64f91e9aae86b3"} Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.257908 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" event={"ID":"fc341ad8-861e-427c-835c-b7bd7367b4a0","Type":"ContainerStarted","Data":"3b9a1ba42d31b3c723a88647a97bc4ce61e95f598beee40d1ce8c2eb4b869ed8"} Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.260108 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-n8lb6" event={"ID":"f9e9c612-02e2-4817-920f-e500720bd6ff","Type":"ContainerStarted","Data":"07b3ceeb2a1208b5e34e38c2be9f81df6d040deebdeccc979be2aa7f3fdb9662"} Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.263967 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.264105 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.764085006 +0000 UTC m=+149.406105576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.264288 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.264618 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.764604742 +0000 UTC m=+149.406625302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.365553 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.365730 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.865705145 +0000 UTC m=+149.507725705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.365826 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.366324 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.866315522 +0000 UTC m=+149.508336082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.467649 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.468126 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:03.968111615 +0000 UTC m=+149.610132175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.568781 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.569154 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.069140096 +0000 UTC m=+149.711160656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.669944 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.670293 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.17027256 +0000 UTC m=+149.812293110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.716127 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7"] Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.732577 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29412000-5584k"] Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.737135 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk"] Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.742665 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99"] Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.747261 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7ckmb"] Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.755154 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b5vr9"] Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.771700 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.772213 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.272199117 +0000 UTC m=+149.914219677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.782791 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd"] Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.790403 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tqgnx"] Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.874167 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.874294 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.374270508 +0000 UTC m=+150.016291068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.874505 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.874794 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.374780853 +0000 UTC m=+150.016801413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:03 crc kubenswrapper[4912]: I1203 00:26:03.975596 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:03 crc kubenswrapper[4912]: E1203 00:26:03.975876 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.475860866 +0000 UTC m=+150.117881426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.015744 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.041578 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.042845 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-cg7ls"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.066368 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kd5n5"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.080151 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.080530 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.580514951 +0000 UTC m=+150.222535511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.094794 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.098629 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-bbm58"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.126494 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.132989 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7x9ms"] Dec 03 00:26:04 crc kubenswrapper[4912]: W1203 00:26:04.143146 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac6ae3d3_462a_4a82_a4a6_e225f12ce343.slice/crio-840fd9307b115ccd2ab5a463816a2a3dcc04750cbcf50b217cfd58e68cbd7888 WatchSource:0}: Error finding container 840fd9307b115ccd2ab5a463816a2a3dcc04750cbcf50b217cfd58e68cbd7888: Status 404 returned error can't find the container with id 840fd9307b115ccd2ab5a463816a2a3dcc04750cbcf50b217cfd58e68cbd7888 Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.143737 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v"] Dec 03 00:26:04 crc kubenswrapper[4912]: W1203 00:26:04.143969 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-969f67d58bf537ec454ea0a9e7529da87febe88bf2ee59b59fac2b7c9c0fa166 WatchSource:0}: Error finding container 969f67d58bf537ec454ea0a9e7529da87febe88bf2ee59b59fac2b7c9c0fa166: Status 404 returned error can't find the container with id 969f67d58bf537ec454ea0a9e7529da87febe88bf2ee59b59fac2b7c9c0fa166 Dec 03 00:26:04 crc kubenswrapper[4912]: W1203 00:26:04.145537 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc0875cd_1663_45f8_a70b_96b0fb672d08.slice/crio-55032e5107151a673c91f6f6c74cf696fc239e842014c7a62636634a0d90a85f WatchSource:0}: Error finding container 55032e5107151a673c91f6f6c74cf696fc239e842014c7a62636634a0d90a85f: Status 404 returned error can't find the container with id 55032e5107151a673c91f6f6c74cf696fc239e842014c7a62636634a0d90a85f Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.147221 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4pwws"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.151141 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.152636 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-t69lf"] Dec 03 00:26:04 crc kubenswrapper[4912]: W1203 00:26:04.157554 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-f65a584dde1165f4218566bfa9e0b49fd67fc3eb79469924c410144b2915e485 WatchSource:0}: Error finding container f65a584dde1165f4218566bfa9e0b49fd67fc3eb79469924c410144b2915e485: Status 404 returned error can't find the container with id f65a584dde1165f4218566bfa9e0b49fd67fc3eb79469924c410144b2915e485 Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.157962 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lg5fq"] Dec 03 00:26:04 crc kubenswrapper[4912]: W1203 00:26:04.158698 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfaa0cccb_7352_4752_8957_0d770dacaee5.slice/crio-7c010b59ce3b3661f1e87ec3050603cbc3f7e608d37f407cf1e5ce4ee0b7912d WatchSource:0}: Error finding container 7c010b59ce3b3661f1e87ec3050603cbc3f7e608d37f407cf1e5ce4ee0b7912d: Status 404 returned error can't find the container with id 7c010b59ce3b3661f1e87ec3050603cbc3f7e608d37f407cf1e5ce4ee0b7912d Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.190666 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.191349 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.691332936 +0000 UTC m=+150.333353496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.293129 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.293507 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.7934956 +0000 UTC m=+150.435516160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.314668 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" event={"ID":"faa0cccb-7352-4752-8957-0d770dacaee5","Type":"ContainerStarted","Data":"7c010b59ce3b3661f1e87ec3050603cbc3f7e608d37f407cf1e5ce4ee0b7912d"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.335986 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4pwws" event={"ID":"54c92a35-71f5-4fe9-9936-c7a3ea0990c8","Type":"ContainerStarted","Data":"306313ea12baa78a71d1f20e1acf4741da8dd3f1e9d601de65984d41a9b38082"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.357919 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29412000-5584k" event={"ID":"2839435a-a464-473c-94ad-7673ad00630b","Type":"ContainerStarted","Data":"61b58ffc85505a8bb13404bdeba67fd7db900bdd1edad23f8d68de78f3430e86"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.359767 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" event={"ID":"23523b1a-5f26-4f73-8ae9-a9735a83a7de","Type":"ContainerStarted","Data":"fdf351869b052ddcc1f0e6aaed192dc74c195dcbd976bafbf98fad2f2f952e96"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.360757 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.362399 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" event={"ID":"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed","Type":"ContainerStarted","Data":"e0fe20da6fa991d143d9451ba1b53ab718c05d47809e1d8ddd2d08fe28b0101e"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.362899 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.368869 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" event={"ID":"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2","Type":"ContainerStarted","Data":"de8df111169b48e2042727d05227d4f3c24887f6b8cd06c163c111d112a07454"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.370580 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" event={"ID":"bd6c8225-9d04-4ecf-9163-21508a4fd09a","Type":"ContainerStarted","Data":"6c5f27e35e9c8ef99f3541c7b757271d8e56ad76dd5c070e7d6b78e179345c59"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.371727 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" event={"ID":"ac6ae3d3-462a-4a82-a4a6-e225f12ce343","Type":"ContainerStarted","Data":"840fd9307b115ccd2ab5a463816a2a3dcc04750cbcf50b217cfd58e68cbd7888"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.373054 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b5vr9" event={"ID":"61f377a2-06fc-427d-9c3e-38f9b3473e23","Type":"ContainerStarted","Data":"6efa4c8bd70e0cd998b5b1bbf68fd8eee2b13016c6d4deb9866eb206ef17cedb"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.375796 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v" event={"ID":"cdbf7d39-1149-49d5-956b-0d2d31362c0e","Type":"ContainerStarted","Data":"71c56148066a793616d02e8afd64fc1b506a89c377c5ac27556b4631140b0371"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.376732 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" event={"ID":"e8990726-b55a-4424-8d40-18132fbc955c","Type":"ContainerStarted","Data":"8e03c0e7b06de19a1f0e5938fc0e5b5afc37be3ba6d3316f1ae24550fe55d17b"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.377531 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" event={"ID":"d00bf964-db4a-449d-ad97-d563b4b70f50","Type":"ContainerStarted","Data":"90e2d8ed356b3c3b61e7b5470014839fd0a78a8c281ab93e36d4e0e89d42b5c7"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.380448 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" event={"ID":"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a","Type":"ContainerStarted","Data":"d19a26caf6b142a3c4ba02e7d09d79f0cdd32d5e4a8e6abd3983142e0dacc1dc"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.381270 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" event={"ID":"f33e67e7-af60-4c02-86f0-1bfd3de3d690","Type":"ContainerStarted","Data":"db2a63aea8194bf1e4ab0a2a919e87cb2c6211ccf184a05791180f7830c2a677"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.382370 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" event={"ID":"59f8296f-ab2b-40bc-af9c-80a1102321fa","Type":"ContainerStarted","Data":"bcb6052dc2f4a3ce689669f083355728d63bf024d0d527e327d83890af705f7e"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.383231 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" event={"ID":"50003e5f-f76f-486a-bb7d-70a4e553e0cf","Type":"ContainerStarted","Data":"e3c479cc541e58d5fcde4e1be2cb4e847033f1f296e12e5ca77759d9ec698599"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.384240 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" event={"ID":"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0","Type":"ContainerStarted","Data":"4ca821fce3b959573153d9369a93ee7c041027dcdc330cda53d52524cb8e4802"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.385258 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"969f67d58bf537ec454ea0a9e7529da87febe88bf2ee59b59fac2b7c9c0fa166"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.386141 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" event={"ID":"84edc937-d48f-439b-90a7-df4917ead6eb","Type":"ContainerStarted","Data":"96a2f078d4d3db7f3adf75089091ced4358b0c88f8183b919ed9b7a4bc65cbce"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.386914 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" event={"ID":"dc0875cd-1663-45f8-a70b-96b0fb672d08","Type":"ContainerStarted","Data":"55032e5107151a673c91f6f6c74cf696fc239e842014c7a62636634a0d90a85f"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.388046 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f65a584dde1165f4218566bfa9e0b49fd67fc3eb79469924c410144b2915e485"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.388976 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cg7ls" event={"ID":"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93","Type":"ContainerStarted","Data":"3ced7fb72c9da7aafb7b0b7e960dd14acfb8a96219f5081e6832527e331b99dc"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.389786 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" event={"ID":"38ad0fca-841e-4f4f-a6b3-ad172aa5b781","Type":"ContainerStarted","Data":"5d6ebad0302374dfb7967a2b34132bc95227436d4c13750530834e7fd3f0875f"} Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.394567 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.394920 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.894906292 +0000 UTC m=+150.536926852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.495996 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.496343 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:04.996327924 +0000 UTC m=+150.638348484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.598102 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.598740 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.098717725 +0000 UTC m=+150.740738295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.687940 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vdwv6"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.691336 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.692265 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgskz"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.694576 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.699730 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.701485 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-gqt86"] Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.701783 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.201766414 +0000 UTC m=+150.843786974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.703277 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.707199 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.734124 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-c65h9"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.745866 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.769395 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.792968 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-njltp"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.803223 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.803588 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.303537167 +0000 UTC m=+150.945557737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.803964 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.804634 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.304616558 +0000 UTC m=+150.946637118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.806513 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.826725 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4z6ww"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.846484 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vldb5"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.863054 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mv4s4"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.887891 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k"] Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.905025 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.905185 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.405164454 +0000 UTC m=+151.047185024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:04 crc kubenswrapper[4912]: I1203 00:26:04.905271 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:04 crc kubenswrapper[4912]: E1203 00:26:04.905547 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.405539506 +0000 UTC m=+151.047560066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.006858 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.007156 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.507141683 +0000 UTC m=+151.149162243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: W1203 00:26:05.066731 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd032b60_199f_4f51_b635_11666dc68b7c.slice/crio-ab05ce73bbc5f2578ace0609e5a697a2d8a59fef6f64442d5f942f0b92bbedde WatchSource:0}: Error finding container ab05ce73bbc5f2578ace0609e5a697a2d8a59fef6f64442d5f942f0b92bbedde: Status 404 returned error can't find the container with id ab05ce73bbc5f2578ace0609e5a697a2d8a59fef6f64442d5f942f0b92bbedde Dec 03 00:26:05 crc kubenswrapper[4912]: W1203 00:26:05.068923 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8054629_8c7a_4929_92c4_f720e3d30872.slice/crio-432a2f3bde108208ceb863c091a5a155b6503bf2912cc1760a83dc81f7398560 WatchSource:0}: Error finding container 432a2f3bde108208ceb863c091a5a155b6503bf2912cc1760a83dc81f7398560: Status 404 returned error can't find the container with id 432a2f3bde108208ceb863c091a5a155b6503bf2912cc1760a83dc81f7398560 Dec 03 00:26:05 crc kubenswrapper[4912]: W1203 00:26:05.095742 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc916e93d_66d8_436b_841e_c7a92f823912.slice/crio-135405967aa9740987410dff7772d719cef83a444ba490724d8d97d7f0a93cdd WatchSource:0}: Error finding container 135405967aa9740987410dff7772d719cef83a444ba490724d8d97d7f0a93cdd: Status 404 returned error can't find the container with id 135405967aa9740987410dff7772d719cef83a444ba490724d8d97d7f0a93cdd Dec 03 00:26:05 crc kubenswrapper[4912]: W1203 00:26:05.096526 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dfca77d_06b5_41dc_bbf6_4cf48b56e9c9.slice/crio-4acbe6b9e791d959da24ca38f0db015f32d8e534c8b035fc4a0e469fea917824 WatchSource:0}: Error finding container 4acbe6b9e791d959da24ca38f0db015f32d8e534c8b035fc4a0e469fea917824: Status 404 returned error can't find the container with id 4acbe6b9e791d959da24ca38f0db015f32d8e534c8b035fc4a0e469fea917824 Dec 03 00:26:05 crc kubenswrapper[4912]: W1203 00:26:05.098040 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcff6f92d_904b_43ba_a0d7_2bae917cdcd8.slice/crio-cbafe0f2d97a7ac8871836782b3e28bfca7cf8cda471bb9fa14972c4d6c9023a WatchSource:0}: Error finding container cbafe0f2d97a7ac8871836782b3e28bfca7cf8cda471bb9fa14972c4d6c9023a: Status 404 returned error can't find the container with id cbafe0f2d97a7ac8871836782b3e28bfca7cf8cda471bb9fa14972c4d6c9023a Dec 03 00:26:05 crc kubenswrapper[4912]: W1203 00:26:05.098360 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbff567eb_a56f_42e6_9dc2_a58d8b2d8df0.slice/crio-77b43012dac4ba8ff623bed72b86d51a3446303ab7d9767fd93a956909e1a88c WatchSource:0}: Error finding container 77b43012dac4ba8ff623bed72b86d51a3446303ab7d9767fd93a956909e1a88c: Status 404 returned error can't find the container with id 77b43012dac4ba8ff623bed72b86d51a3446303ab7d9767fd93a956909e1a88c Dec 03 00:26:05 crc kubenswrapper[4912]: W1203 00:26:05.099230 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod975f5428_8720_4c72_8595_f213ef97888e.slice/crio-28bd7ee4c4f80167c648a75f018b0b3dd52c46121fba03e9e69ba8eefbd995dc WatchSource:0}: Error finding container 28bd7ee4c4f80167c648a75f018b0b3dd52c46121fba03e9e69ba8eefbd995dc: Status 404 returned error can't find the container with id 28bd7ee4c4f80167c648a75f018b0b3dd52c46121fba03e9e69ba8eefbd995dc Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.108043 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.108302 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.608290367 +0000 UTC m=+151.250310927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.209800 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.210122 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.710105171 +0000 UTC m=+151.352125731 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: W1203 00:26:05.252000 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94036057_869a_49f0_84ca_8397bab8472c.slice/crio-fe72e082972cef90cd6948641a38514867827e25f95edfc53dd5f73538a7f7fe WatchSource:0}: Error finding container fe72e082972cef90cd6948641a38514867827e25f95edfc53dd5f73538a7f7fe: Status 404 returned error can't find the container with id fe72e082972cef90cd6948641a38514867827e25f95edfc53dd5f73538a7f7fe Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.312650 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.315907 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.815846509 +0000 UTC m=+151.457867069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.398583 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vldb5" event={"ID":"82f9b623-024f-45e6-b876-f9718bc8aba6","Type":"ContainerStarted","Data":"467ba5f5e7a22b8e1a3a88562859d5284327e92ed844e226aadbd354f43fbb2b"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.399745 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" event={"ID":"cff6f92d-904b-43ba-a0d7-2bae917cdcd8","Type":"ContainerStarted","Data":"cbafe0f2d97a7ac8871836782b3e28bfca7cf8cda471bb9fa14972c4d6c9023a"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.404887 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" event={"ID":"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0","Type":"ContainerStarted","Data":"dfa12b3a9f1351781b9771685e9ac95e97dd1874e64049fc2392e517090e04b3"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.408992 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" event={"ID":"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a","Type":"ContainerStarted","Data":"27e7499ea05b4dfb3d9901f2dd44c3c02d9f67f3a5d1b483c68c516656816aa1"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.412871 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" event={"ID":"fc341ad8-861e-427c-835c-b7bd7367b4a0","Type":"ContainerStarted","Data":"64c9d32b4ad1ef7fec75962b1da81f39dbcd264bc30ac60f56b6669b20260cdd"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.414656 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.414753 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.914737958 +0000 UTC m=+151.556758518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.414982 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.415246 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:05.915239803 +0000 UTC m=+151.557260353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.415344 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" event={"ID":"d00bf964-db4a-449d-ad97-d563b4b70f50","Type":"ContainerStarted","Data":"a99002b2a1d9a859d66bdc1b157277ad9fe2c114d9fbf1cb5d519cd3eeb2506f"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.418518 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-njltp" event={"ID":"0f44254a-d149-4a4b-aec0-1a2613d47f81","Type":"ContainerStarted","Data":"602cb01450cd2c0c53bcc7fbd99548c811e9fac24d7cf9a22133be6f544caa5c"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.419755 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" event={"ID":"edb6b379-aeff-470a-a09d-540d6acc0efe","Type":"ContainerStarted","Data":"1daa0b3adab8e5821a74767d7e0fca5d3ffd55c42f694f801a825f38dec3697f"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.421247 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" event={"ID":"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679","Type":"ContainerStarted","Data":"87e1d7b414dd99ca62a2b4fd3ff85474c3722aab564fbb4b5cdffae5a551d95c"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.423251 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" event={"ID":"a8054629-8c7a-4929-92c4-f720e3d30872","Type":"ContainerStarted","Data":"432a2f3bde108208ceb863c091a5a155b6503bf2912cc1760a83dc81f7398560"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.425006 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" event={"ID":"c916e93d-66d8-436b-841e-c7a92f823912","Type":"ContainerStarted","Data":"135405967aa9740987410dff7772d719cef83a444ba490724d8d97d7f0a93cdd"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.430973 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" event={"ID":"e8990726-b55a-4424-8d40-18132fbc955c","Type":"ContainerStarted","Data":"6076b4cd7d3afac85222163d7933cce213bdaead8b4916f14d143b859dc7679f"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.432920 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-gqt86" event={"ID":"9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9","Type":"ContainerStarted","Data":"4acbe6b9e791d959da24ca38f0db015f32d8e534c8b035fc4a0e469fea917824"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.434459 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" event={"ID":"f6a647db-07ba-49f6-9150-0a038d4fbee8","Type":"ContainerStarted","Data":"5bb77a65814e7b8ebcb4549e641f2eeaf19131e6ab981d5e6fcdd4008699af09"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.435610 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" event={"ID":"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8","Type":"ContainerStarted","Data":"5babb46b07ff49ea339d92679d3e4e71bbe9e43fa5fe38da15004e27be35610f"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.437289 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b5vr9" event={"ID":"61f377a2-06fc-427d-9c3e-38f9b3473e23","Type":"ContainerStarted","Data":"0e2f9a41224c4e26226217fa58b4a1c0d8dc10266e6cb15fe5b07b2309076ecc"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.438348 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" event={"ID":"94036057-869a-49f0-84ca-8397bab8472c","Type":"ContainerStarted","Data":"fe72e082972cef90cd6948641a38514867827e25f95edfc53dd5f73538a7f7fe"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.440041 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" event={"ID":"ac6ae3d3-462a-4a82-a4a6-e225f12ce343","Type":"ContainerStarted","Data":"b301cc27f2e057233af33e1feb1dfaa5aa6c56907b42bd37185551945b2cbcf2"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.442833 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-n8lb6" event={"ID":"f9e9c612-02e2-4817-920f-e500720bd6ff","Type":"ContainerStarted","Data":"ff90fd2b895a5ff5a3f549c83d22ec9b8efe20433db830b73af4759a393eb2df"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.444571 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" event={"ID":"28f13442-6621-41b8-9d00-b6607702fa48","Type":"ContainerStarted","Data":"9ad8403deaa85f900a324eb9e160a5b9ab696917e7f01f109f6b2c0db7c6d622"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.445835 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" event={"ID":"896d21a9-3b77-4571-a4c2-edc02fdf6db2","Type":"ContainerStarted","Data":"2bef8afe9ef8240c1fc951e12912c909f020bc597db58d4f20f6a18b6278a88e"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.446913 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" event={"ID":"8f8b45be-2bdf-447c-82f7-758871f840e3","Type":"ContainerStarted","Data":"9ac3508f1fa1b7c5128ab5d1436ba26b795b84c97c8edd46bd715818ecaf0268"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.447667 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" event={"ID":"50d1bf1c-014f-4900-a835-96cf690299e9","Type":"ContainerStarted","Data":"113a040f9eae671f1cf995de5856ff6fbd4daa81354dbe65b7bbec9c4a40b139"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.448708 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" event={"ID":"84edc937-d48f-439b-90a7-df4917ead6eb","Type":"ContainerStarted","Data":"bdeb3fb333371c609766353e301670343ae66f6a9880f96fad16a32a563fbd9f"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.449553 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vdwv6" event={"ID":"fd032b60-199f-4f51-b635-11666dc68b7c","Type":"ContainerStarted","Data":"ab05ce73bbc5f2578ace0609e5a697a2d8a59fef6f64442d5f942f0b92bbedde"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.451811 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" event={"ID":"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0","Type":"ContainerStarted","Data":"77b43012dac4ba8ff623bed72b86d51a3446303ab7d9767fd93a956909e1a88c"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.455340 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29412000-5584k" event={"ID":"2839435a-a464-473c-94ad-7673ad00630b","Type":"ContainerStarted","Data":"1a987a1607172aa64b660afd6ce811b1aefc7dbbd59a4da36e6868ff179de61a"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.460477 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" event={"ID":"dc0875cd-1663-45f8-a70b-96b0fb672d08","Type":"ContainerStarted","Data":"3afc0f9fefb5f8649784da79fc6de8e3d50a51a96c97551a4d63e2fefd72d54d"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.462861 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" event={"ID":"975f5428-8720-4c72-8595-f213ef97888e","Type":"ContainerStarted","Data":"28bd7ee4c4f80167c648a75f018b0b3dd52c46121fba03e9e69ba8eefbd995dc"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.464727 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cg7ls" event={"ID":"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93","Type":"ContainerStarted","Data":"fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed"} Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.485995 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-dddxf" podStartSLOduration=131.485968247 podStartE2EDuration="2m11.485968247s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:05.483261449 +0000 UTC m=+151.125282019" watchObservedRunningTime="2025-12-03 00:26:05.485968247 +0000 UTC m=+151.127988807" Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.517882 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.518471 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.018393925 +0000 UTC m=+151.660414635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.619267 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.619658 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.119641493 +0000 UTC m=+151.761662053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.720565 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.720694 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.220665584 +0000 UTC m=+151.862686144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.721374 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.721738 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.221724275 +0000 UTC m=+151.863744835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.823646 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.824005 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.323984871 +0000 UTC m=+151.966005431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.902521 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.912265 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:05 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:05 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:05 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.912320 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:05 crc kubenswrapper[4912]: I1203 00:26:05.929517 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:05 crc kubenswrapper[4912]: E1203 00:26:05.930179 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.43016692 +0000 UTC m=+152.072187470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.030613 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.030834 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.53080475 +0000 UTC m=+152.172825360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.030958 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.031318 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.531308595 +0000 UTC m=+152.173329155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.135389 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.135751 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.635731184 +0000 UTC m=+152.277751744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.236881 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.237201 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.737189868 +0000 UTC m=+152.379210418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.338150 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.338588 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.838570769 +0000 UTC m=+152.480591329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.440278 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.442080 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:06.942061671 +0000 UTC m=+152.584082231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.543203 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.543627 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:07.043610718 +0000 UTC m=+152.685631278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.567899 4912 generic.go:334] "Generic (PLEG): container finished" podID="3c4ce8f2-b6c1-4603-91b7-47e38cd693c2" containerID="289b52b087113dd0451de0273a3265a0393272d353521b4789f29ee6e1a1220e" exitCode=0 Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.567953 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" event={"ID":"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2","Type":"ContainerDied","Data":"289b52b087113dd0451de0273a3265a0393272d353521b4789f29ee6e1a1220e"} Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.647194 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" event={"ID":"f33e67e7-af60-4c02-86f0-1bfd3de3d690","Type":"ContainerStarted","Data":"febe37126fabf8ec245b1ec2adf967cc43b2ae15a06c8444cef431b5bac71e6b"} Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.649223 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.657274 4912 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-7x9ms container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.657362 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" podUID="f33e67e7-af60-4c02-86f0-1bfd3de3d690" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.658735 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.740588 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:07.240560802 +0000 UTC m=+152.882581362 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.770781 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" event={"ID":"50d1bf1c-014f-4900-a835-96cf690299e9","Type":"ContainerStarted","Data":"d07ea9f4a755c4c0818ef280c175484141e5db26265eefcb4dce385f8afd4850"} Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.771380 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.773894 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.774617 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:07.274602667 +0000 UTC m=+152.916623227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.785419 4912 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-gm5mt container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.785512 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" podUID="50d1bf1c-014f-4900-a835-96cf690299e9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.803404 4912 generic.go:334] "Generic (PLEG): container finished" podID="9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0" containerID="dfa12b3a9f1351781b9771685e9ac95e97dd1874e64049fc2392e517090e04b3" exitCode=0 Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.807783 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" event={"ID":"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0","Type":"ContainerDied","Data":"dfa12b3a9f1351781b9771685e9ac95e97dd1874e64049fc2392e517090e04b3"} Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.827884 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" podStartSLOduration=132.827864967 podStartE2EDuration="2m12.827864967s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:06.827154086 +0000 UTC m=+152.469174646" watchObservedRunningTime="2025-12-03 00:26:06.827864967 +0000 UTC m=+152.469885527" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.829669 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" podStartSLOduration=132.829657878 podStartE2EDuration="2m12.829657878s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:06.741348605 +0000 UTC m=+152.383369185" watchObservedRunningTime="2025-12-03 00:26:06.829657878 +0000 UTC m=+152.471678438" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.882537 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.884495 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:07.384475143 +0000 UTC m=+153.026495763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.893697 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d4116647bce0e084537e972690f0df7f26ff8f9236e40b4bbc30ec289065314a"} Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.894636 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.906361 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:06 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:06 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:06 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.906464 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.947034 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" event={"ID":"faa0cccb-7352-4752-8957-0d770dacaee5","Type":"ContainerStarted","Data":"161ead0b788d10b8512b924aca6a2951e82533abc1a25202e850c2e717a2f0ea"} Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.947511 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.963340 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" event={"ID":"c916e93d-66d8-436b-841e-c7a92f823912","Type":"ContainerStarted","Data":"acb670445e0d86c859c31d49fe0734cd7907ebf6d9586c532ee7efaea4e0c711"} Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.964495 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.983762 4912 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dgskz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.984304 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.984440 4912 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-kd5n5 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" start-of-body= Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.984462 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" podUID="faa0cccb-7352-4752-8957-0d770dacaee5" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.22:6443/healthz\": dial tcp 10.217.0.22:6443: connect: connection refused" Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.985295 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:06 crc kubenswrapper[4912]: E1203 00:26:06.985991 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:07.485962998 +0000 UTC m=+153.127983568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:06 crc kubenswrapper[4912]: I1203 00:26:06.994403 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vdwv6" event={"ID":"fd032b60-199f-4f51-b635-11666dc68b7c","Type":"ContainerStarted","Data":"f0753f9845f7e4545b065c62c835c3130c9255c1e9527a3a412c64227740ad62"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.025138 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" event={"ID":"50003e5f-f76f-486a-bb7d-70a4e553e0cf","Type":"ContainerStarted","Data":"705c2842b5824d80b57c87f4a5f8b68ed5e3644301c345831b069dc5743d6681"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.056800 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" event={"ID":"23523b1a-5f26-4f73-8ae9-a9735a83a7de","Type":"ContainerStarted","Data":"8bd77884666e1f4d0647f384b61d9442bfecd9b825dc739525d6ba238a179e34"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.057810 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.077306 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" podStartSLOduration=134.077286068 podStartE2EDuration="2m14.077286068s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:06.985138903 +0000 UTC m=+152.627159483" watchObservedRunningTime="2025-12-03 00:26:07.077286068 +0000 UTC m=+152.719306628" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.091765 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" event={"ID":"38ad0fca-841e-4f4f-a6b3-ad172aa5b781","Type":"ContainerStarted","Data":"0f934c64ae5c47d7ca46a2d916039937e1e20860a084bc173b1368ba84c4f473"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.094058 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:07 crc kubenswrapper[4912]: E1203 00:26:07.101057 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:07.601042345 +0000 UTC m=+153.243062905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.131138 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" event={"ID":"59f8296f-ab2b-40bc-af9c-80a1102321fa","Type":"ContainerStarted","Data":"8acfcd560317dd70a578444f299318c6c7fc86e5ea56b9a44cd039a6e79911f1"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.134713 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.155168 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" podStartSLOduration=133.155143778 podStartE2EDuration="2m13.155143778s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.078130752 +0000 UTC m=+152.720151312" watchObservedRunningTime="2025-12-03 00:26:07.155143778 +0000 UTC m=+152.797164348" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.167628 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" event={"ID":"f6a647db-07ba-49f6-9150-0a038d4fbee8","Type":"ContainerStarted","Data":"40ce820d269d557ca9832e26779b1fc816bf349512285458c543914afb6d317b"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.195063 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:07 crc kubenswrapper[4912]: E1203 00:26:07.197291 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:07.697264636 +0000 UTC m=+153.339285196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.210675 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7ckmb" podStartSLOduration=133.210656504 podStartE2EDuration="2m13.210656504s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.163790099 +0000 UTC m=+152.805810669" watchObservedRunningTime="2025-12-03 00:26:07.210656504 +0000 UTC m=+152.852677064" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.266952 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4pwws" event={"ID":"54c92a35-71f5-4fe9-9936-c7a3ea0990c8","Type":"ContainerStarted","Data":"1e2c221a46e8c2afd46928a889837f098ddb04abbaa077d553412bc3ae132560"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.275468 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.284453 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-bbm58" podStartSLOduration=132.284421347 podStartE2EDuration="2m12.284421347s" podCreationTimestamp="2025-12-03 00:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.282874852 +0000 UTC m=+152.924895412" watchObservedRunningTime="2025-12-03 00:26:07.284421347 +0000 UTC m=+152.926441907" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.284596 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-825zz" podStartSLOduration=133.284592052 podStartE2EDuration="2m13.284592052s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.214636089 +0000 UTC m=+152.856656649" watchObservedRunningTime="2025-12-03 00:26:07.284592052 +0000 UTC m=+152.926612612" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.294048 4912 patch_prober.go:28] interesting pod/console-operator-58897d9998-4pwws container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.294099 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4pwws" podUID="54c92a35-71f5-4fe9-9936-c7a3ea0990c8" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/readyz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.296201 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:07 crc kubenswrapper[4912]: E1203 00:26:07.296488 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:07.796472615 +0000 UTC m=+153.438493175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.314057 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v" event={"ID":"cdbf7d39-1149-49d5-956b-0d2d31362c0e","Type":"ContainerStarted","Data":"465ecb4530d37707a7592caae1d61984db7b8c3b21f78aa1ee93cdb97741cedc"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.315843 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" podStartSLOduration=134.315833155 podStartE2EDuration="2m14.315833155s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.314369793 +0000 UTC m=+152.956390343" watchObservedRunningTime="2025-12-03 00:26:07.315833155 +0000 UTC m=+152.957853715" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.325843 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" event={"ID":"ec9265e2-f966-4b1e-bd5d-22ed1c4fc6e8","Type":"ContainerStarted","Data":"b3f2814415248c8dda8dc7bbff622ad191767260cdeb93821877b41e0675baf2"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.339635 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" event={"ID":"8f8b45be-2bdf-447c-82f7-758871f840e3","Type":"ContainerStarted","Data":"f2053d3a2c845f4850d4acaece71ae2d6312bbbfc8c355077ddfd0c9810a2731"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.351713 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7d7edc25970240cdb5afda318ffeb891036b6b51d69f6b00871eabd190c80305"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.397349 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-4pwws" podStartSLOduration=133.397331761 podStartE2EDuration="2m13.397331761s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.39483633 +0000 UTC m=+153.036856890" watchObservedRunningTime="2025-12-03 00:26:07.397331761 +0000 UTC m=+153.039352321" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.398501 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:07 crc kubenswrapper[4912]: E1203 00:26:07.400518 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:07.900474463 +0000 UTC m=+153.542495043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.415867 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" event={"ID":"ea1e4d09-7fd1-4bcf-b7f7-574ccaba08ed","Type":"ContainerStarted","Data":"bf6907bb550fa5c2976f5aa9aef782ce5c1f13d34068f4fcf4447a93ccdc5d84"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.457113 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-4z6ww" podStartSLOduration=134.456912614 podStartE2EDuration="2m14.456912614s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.442667182 +0000 UTC m=+153.084687742" watchObservedRunningTime="2025-12-03 00:26:07.456912614 +0000 UTC m=+153.098933174" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.458239 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" event={"ID":"a1e2de7d-cdd8-433b-b42c-57e6cb6d8679","Type":"ContainerStarted","Data":"a63f69305b475700da8967d678d7abfdfb27e0444035340e11be82c6711aefd6"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.461863 4912 generic.go:334] "Generic (PLEG): container finished" podID="975f5428-8720-4c72-8595-f213ef97888e" containerID="63ced86ff398cefb63002d7e434d40411330ab9e4bd7838aeb18cf2eaa93b80b" exitCode=0 Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.461910 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" event={"ID":"975f5428-8720-4c72-8595-f213ef97888e","Type":"ContainerDied","Data":"63ced86ff398cefb63002d7e434d40411330ab9e4bd7838aeb18cf2eaa93b80b"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.501291 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.502701 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" event={"ID":"bd6c8225-9d04-4ecf-9163-21508a4fd09a","Type":"ContainerStarted","Data":"6d3b3c45bc337ea8c3a22e8233ff44a9982522b97eff64ebd9aa8bb693bf2404"} Dec 03 00:26:07 crc kubenswrapper[4912]: E1203 00:26:07.505281 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.005267402 +0000 UTC m=+153.647288012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.510185 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-gqt86" event={"ID":"9dfca77d-06b5-41dc-bbf6-4cf48b56e9c9","Type":"ContainerStarted","Data":"e0f7295a246a1fc0091160ae7413b761dfb345c0669ec854ebd8736077ed7db3"} Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.513420 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.530644 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.576148 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-d9mkw" podStartSLOduration=133.576125401 podStartE2EDuration="2m13.576125401s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.502346937 +0000 UTC m=+153.144367497" watchObservedRunningTime="2025-12-03 00:26:07.576125401 +0000 UTC m=+153.218145951" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.634096 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-mv4s4" podStartSLOduration=132.634077936 podStartE2EDuration="2m12.634077936s" podCreationTimestamp="2025-12-03 00:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.592983298 +0000 UTC m=+153.235003878" watchObservedRunningTime="2025-12-03 00:26:07.634077936 +0000 UTC m=+153.276098496" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.636808 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:07 crc kubenswrapper[4912]: E1203 00:26:07.639691 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.139658777 +0000 UTC m=+153.781679337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.703450 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zmv99" podStartSLOduration=134.703408601 podStartE2EDuration="2m14.703408601s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.692117045 +0000 UTC m=+153.334137615" watchObservedRunningTime="2025-12-03 00:26:07.703408601 +0000 UTC m=+153.345429161" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.740876 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.741102 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-b5vr9" podStartSLOduration=133.74108143 podStartE2EDuration="2m13.74108143s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.739884816 +0000 UTC m=+153.381905386" watchObservedRunningTime="2025-12-03 00:26:07.74108143 +0000 UTC m=+153.383101990" Dec 03 00:26:07 crc kubenswrapper[4912]: E1203 00:26:07.752110 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.252084879 +0000 UTC m=+153.894105439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.806412 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-5rhpw" podStartSLOduration=133.806391408 podStartE2EDuration="2m13.806391408s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.791019735 +0000 UTC m=+153.433040305" watchObservedRunningTime="2025-12-03 00:26:07.806391408 +0000 UTC m=+153.448411978" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.841149 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xf4wl" podStartSLOduration=133.841131353 podStartE2EDuration="2m13.841131353s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.840988649 +0000 UTC m=+153.483009219" watchObservedRunningTime="2025-12-03 00:26:07.841131353 +0000 UTC m=+153.483151913" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.841566 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:07 crc kubenswrapper[4912]: E1203 00:26:07.842049 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.342029849 +0000 UTC m=+153.984050409 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.913655 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" podStartSLOduration=133.913637499 podStartE2EDuration="2m13.913637499s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.912824286 +0000 UTC m=+153.554844856" watchObservedRunningTime="2025-12-03 00:26:07.913637499 +0000 UTC m=+153.555658059" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.913917 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-gqt86" podStartSLOduration=9.913909707 podStartE2EDuration="9.913909707s" podCreationTimestamp="2025-12-03 00:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.878055801 +0000 UTC m=+153.520076361" watchObservedRunningTime="2025-12-03 00:26:07.913909707 +0000 UTC m=+153.555930267" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.922261 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:07 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:07 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:07 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.922327 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.943317 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:07 crc kubenswrapper[4912]: E1203 00:26:07.943783 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.443768791 +0000 UTC m=+154.085789351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.974681 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-x4gmd" podStartSLOduration=134.974663704 podStartE2EDuration="2m14.974663704s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.954845881 +0000 UTC m=+153.596866461" watchObservedRunningTime="2025-12-03 00:26:07.974663704 +0000 UTC m=+153.616684264" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.975689 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29412000-5584k" podStartSLOduration=134.975681113 podStartE2EDuration="2m14.975681113s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.972861131 +0000 UTC m=+153.614881681" watchObservedRunningTime="2025-12-03 00:26:07.975681113 +0000 UTC m=+153.617701673" Dec 03 00:26:07 crc kubenswrapper[4912]: I1203 00:26:07.995239 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6cgnd" podStartSLOduration=133.995209498 podStartE2EDuration="2m13.995209498s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:07.994860968 +0000 UTC m=+153.636881528" watchObservedRunningTime="2025-12-03 00:26:07.995209498 +0000 UTC m=+153.637230058" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.023614 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-n8lb6" podStartSLOduration=10.023572868 podStartE2EDuration="10.023572868s" podCreationTimestamp="2025-12-03 00:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:08.022397724 +0000 UTC m=+153.664418294" watchObservedRunningTime="2025-12-03 00:26:08.023572868 +0000 UTC m=+153.665593428" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.057058 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.057652 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.557631273 +0000 UTC m=+154.199651843 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.161802 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.162115 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.662101894 +0000 UTC m=+154.304122454 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.266572 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.267370 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.767349437 +0000 UTC m=+154.409369997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.368125 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.368449 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.868421229 +0000 UTC m=+154.510441789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.486944 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.487286 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.987259395 +0000 UTC m=+154.629279945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.487377 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.487738 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:08.987721338 +0000 UTC m=+154.629741898 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.581699 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v" event={"ID":"cdbf7d39-1149-49d5-956b-0d2d31362c0e","Type":"ContainerStarted","Data":"3bc010da7755fab332608ef0d4c3aa6d41c1dd343c65da589250e222e5307702"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.588338 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.588839 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.088816571 +0000 UTC m=+154.730837131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.591085 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vdwv6" event={"ID":"fd032b60-199f-4f51-b635-11666dc68b7c","Type":"ContainerStarted","Data":"9395bc05aa333c712eecb1daed071cef83acc11974a534a2212a904f8ef758cd"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.591923 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.611133 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" event={"ID":"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0","Type":"ContainerStarted","Data":"95a729f4447191e3982fd21b04f01759c3d7a9e59dc16762587844741bfd0b71"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.611195 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" event={"ID":"bff567eb-a56f-42e6-9dc2-a58d8b2d8df0","Type":"ContainerStarted","Data":"9bbeb495d969dd12d044ec0d7d3b5a639ef0f600dae4f0c8fd9de56325c7b345"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.624086 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" event={"ID":"59f8296f-ab2b-40bc-af9c-80a1102321fa","Type":"ContainerStarted","Data":"b7fe43898a292fcc3f0968a407ab91bf6b80585e5f02348cdc8895a193c9bcae"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.641114 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" event={"ID":"28f13442-6621-41b8-9d00-b6607702fa48","Type":"ContainerStarted","Data":"c4ce910e718f4718f55b76b13c4a0cd74bd82552574f6dc98f15cf98ff3e78bd"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.641178 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" event={"ID":"28f13442-6621-41b8-9d00-b6607702fa48","Type":"ContainerStarted","Data":"c8852bc4d5306f53b417f8bdd59a6aa2c419205d9edccfe50058469207d1cabd"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.651962 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" event={"ID":"ac6ae3d3-462a-4a82-a4a6-e225f12ce343","Type":"ContainerStarted","Data":"5e95ed1be1d14befcb33f2fd52540623cc70a80a006549e7957e16fa5e0d429d"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.656807 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" event={"ID":"cff6f92d-904b-43ba-a0d7-2bae917cdcd8","Type":"ContainerStarted","Data":"9fed270cccf6fdcfb4dc47ee9ef141a143873fabac5fcb873a403c008541e85d"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.664444 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" event={"ID":"94036057-869a-49f0-84ca-8397bab8472c","Type":"ContainerStarted","Data":"edc7ff0d7fd2dc2e3f46319f3137f0dafc2d0b4bda002236f85fac1f758c1a89"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.667202 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-cg7ls" podStartSLOduration=134.667181677 podStartE2EDuration="2m14.667181677s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:08.060501206 +0000 UTC m=+153.702521766" watchObservedRunningTime="2025-12-03 00:26:08.667181677 +0000 UTC m=+154.309202237" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.667823 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-8cc6v" podStartSLOduration=134.667816255 podStartE2EDuration="2m14.667816255s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:08.658977779 +0000 UTC m=+154.300998369" watchObservedRunningTime="2025-12-03 00:26:08.667816255 +0000 UTC m=+154.309836815" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.681210 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" event={"ID":"e8990726-b55a-4424-8d40-18132fbc955c","Type":"ContainerStarted","Data":"9e4ec0782dabd53ea52c8d6ea724314f524903bb5195f883eccb19e312e22d57"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.693345 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.693674 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.193662312 +0000 UTC m=+154.835682872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.693813 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" event={"ID":"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2","Type":"ContainerStarted","Data":"c25288821baad35d2c9553c104ccd133e489827afa7a12e2c545301cb9b45da9"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.700939 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vldb5" event={"ID":"82f9b623-024f-45e6-b876-f9718bc8aba6","Type":"ContainerStarted","Data":"d8ff0029e066a783eca002de046d43de222421109f5ea3d9ea6ec318c5792a77"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.702821 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" event={"ID":"a8054629-8c7a-4929-92c4-f720e3d30872","Type":"ContainerStarted","Data":"36799ea45d42d1a1c5337f944c7c294db4c58bc41ebbbdff927883ca0e061bad"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.702853 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" event={"ID":"a8054629-8c7a-4929-92c4-f720e3d30872","Type":"ContainerStarted","Data":"fe6f43f0f1196e07fa2775289d3cbe41f5f44db4917647ea277bc0ef8e151525"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.703012 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.704709 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" event={"ID":"9ed1d04a-aa3d-4c27-a811-b06dabb2f7b0","Type":"ContainerStarted","Data":"219c9a34e8bf5c0a28891bac4d638b5d3d360736a0dedac74208b0d7aa228461"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.706230 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" event={"ID":"975f5428-8720-4c72-8595-f213ef97888e","Type":"ContainerStarted","Data":"3aa0058a4eea093a48d43ea582a21feb7f9c1b09f30664f9c894106c7142c969"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.706285 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.718932 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" event={"ID":"3938650f-17d8-4bb7-9fe7-5399eb3c4a5a","Type":"ContainerStarted","Data":"2e1cd4ac22a70caa3f32cee83562e0130f3af44e6fff15a95e6fffb9abe37812"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.725471 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" event={"ID":"896d21a9-3b77-4571-a4c2-edc02fdf6db2","Type":"ContainerStarted","Data":"21871558010f6f2bff2c62c6e8f9fcd620188ea5b0f86803774608c75b1279df"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.727717 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-njltp" event={"ID":"0f44254a-d149-4a4b-aec0-1a2613d47f81","Type":"ContainerStarted","Data":"ab9df0e81257607c09d7db5fc1a45970c8d17fc86b7c10bc233a5862e78cc4c1"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.727761 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-njltp" event={"ID":"0f44254a-d149-4a4b-aec0-1a2613d47f81","Type":"ContainerStarted","Data":"ad378324d90a794ce46a40b081d269709c6576907b5ded80316559d8f36a05c2"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.732003 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" event={"ID":"edb6b379-aeff-470a-a09d-540d6acc0efe","Type":"ContainerStarted","Data":"84c176cba56f9674d1bee484cedd11df976768ef3229ab4a5f0178193a876c87"} Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.732050 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.735119 4912 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dgskz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.742032 4912 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-cns6n container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" start-of-body= Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.742185 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" podUID="edb6b379-aeff-470a-a09d-540d6acc0efe" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.742269 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.776589 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-gm5mt" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.780053 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zzz2x" podStartSLOduration=134.78003406 podStartE2EDuration="2m14.78003406s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:08.770903305 +0000 UTC m=+154.412923865" watchObservedRunningTime="2025-12-03 00:26:08.78003406 +0000 UTC m=+154.422054620" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.794776 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.804272 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.30424795 +0000 UTC m=+154.946268510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.900187 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:08 crc kubenswrapper[4912]: E1203 00:26:08.900569 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.400557394 +0000 UTC m=+155.042577954 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.916909 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:08 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:08 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:08 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.916960 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.919773 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bq8vd" podStartSLOduration=134.9197548 podStartE2EDuration="2m14.9197548s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:08.90971563 +0000 UTC m=+154.551736190" watchObservedRunningTime="2025-12-03 00:26:08.9197548 +0000 UTC m=+154.561775360" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.974222 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:26:08 crc kubenswrapper[4912]: I1203 00:26:08.987056 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-tqgnx" podStartSLOduration=134.987037305 podStartE2EDuration="2m14.987037305s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:08.985247873 +0000 UTC m=+154.627268433" watchObservedRunningTime="2025-12-03 00:26:08.987037305 +0000 UTC m=+154.629057865" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.000859 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.001011 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.500989108 +0000 UTC m=+155.143009678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.001120 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.001454 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.501444951 +0000 UTC m=+155.143465521 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.022498 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vtnw8" podStartSLOduration=135.022479969 podStartE2EDuration="2m15.022479969s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.021759309 +0000 UTC m=+154.663779869" watchObservedRunningTime="2025-12-03 00:26:09.022479969 +0000 UTC m=+154.664500529" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.042615 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-4pwws" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.088209 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-g5h6h" podStartSLOduration=135.088191229 podStartE2EDuration="2m15.088191229s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.085226084 +0000 UTC m=+154.727246644" watchObservedRunningTime="2025-12-03 00:26:09.088191229 +0000 UTC m=+154.730211789" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.101825 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.102080 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.602066731 +0000 UTC m=+155.244087291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.187993 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vdwv6" podStartSLOduration=10.187962114 podStartE2EDuration="10.187962114s" podCreationTimestamp="2025-12-03 00:25:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.14009307 +0000 UTC m=+154.782113660" watchObservedRunningTime="2025-12-03 00:26:09.187962114 +0000 UTC m=+154.829982674" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.202703 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.203122 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.703106372 +0000 UTC m=+155.345126932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.203981 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lqdfv"] Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.205131 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.214868 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.224729 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lqdfv"] Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.290853 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-xf9jw" podStartSLOduration=135.290804958 podStartE2EDuration="2m15.290804958s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.248101893 +0000 UTC m=+154.890122483" watchObservedRunningTime="2025-12-03 00:26:09.290804958 +0000 UTC m=+154.932825518" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.292459 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s2bl7" podStartSLOduration=135.292449415 podStartE2EDuration="2m15.292449415s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.288672736 +0000 UTC m=+154.930693306" watchObservedRunningTime="2025-12-03 00:26:09.292449415 +0000 UTC m=+154.934469975" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.304969 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.305118 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-catalog-content\") pod \"certified-operators-lqdfv\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.305154 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-utilities\") pod \"certified-operators-lqdfv\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.305203 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4szh6\" (UniqueName: \"kubernetes.io/projected/9e57c9b1-dd22-4cc5-ac09-f010c8030170-kube-api-access-4szh6\") pod \"certified-operators-lqdfv\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.305258 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.805215084 +0000 UTC m=+155.447235644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.397849 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-njltp" podStartSLOduration=135.397831622 podStartE2EDuration="2m15.397831622s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.396916815 +0000 UTC m=+155.038937375" watchObservedRunningTime="2025-12-03 00:26:09.397831622 +0000 UTC m=+155.039852182" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.406487 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-catalog-content\") pod \"certified-operators-lqdfv\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.406552 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.406594 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-utilities\") pod \"certified-operators-lqdfv\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.406658 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4szh6\" (UniqueName: \"kubernetes.io/projected/9e57c9b1-dd22-4cc5-ac09-f010c8030170-kube-api-access-4szh6\") pod \"certified-operators-lqdfv\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.422279 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-catalog-content\") pod \"certified-operators-lqdfv\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.422694 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:09.922678881 +0000 UTC m=+155.564699441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.422709 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-utilities\") pod \"certified-operators-lqdfv\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.427462 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6c6m2"] Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.429199 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.438087 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.505364 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4szh6\" (UniqueName: \"kubernetes.io/projected/9e57c9b1-dd22-4cc5-ac09-f010c8030170-kube-api-access-4szh6\") pod \"certified-operators-lqdfv\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.505971 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6c6m2"] Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.506872 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" podStartSLOduration=135.506853304 podStartE2EDuration="2m15.506853304s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.506100352 +0000 UTC m=+155.148120932" watchObservedRunningTime="2025-12-03 00:26:09.506853304 +0000 UTC m=+155.148873864" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.519125 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.524574 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.024543536 +0000 UTC m=+155.666564096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.519728 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-catalog-content\") pod \"community-operators-6c6m2\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.525400 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7gvq\" (UniqueName: \"kubernetes.io/projected/da8155c3-0290-44f3-b1da-e95d2dcee000-kube-api-access-b7gvq\") pod \"community-operators-6c6m2\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.525556 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-utilities\") pod \"community-operators-6c6m2\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.525623 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.525958 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.025944907 +0000 UTC m=+155.667965467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.537149 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.588367 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-5fg2k" podStartSLOduration=135.58834455 podStartE2EDuration="2m15.58834455s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.565680355 +0000 UTC m=+155.207700915" watchObservedRunningTime="2025-12-03 00:26:09.58834455 +0000 UTC m=+155.230365110" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.615973 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rf9m5"] Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.617686 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" podStartSLOduration=135.617667238 podStartE2EDuration="2m15.617667238s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.612520309 +0000 UTC m=+155.254540869" watchObservedRunningTime="2025-12-03 00:26:09.617667238 +0000 UTC m=+155.259687798" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.617737 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.629740 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.630027 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-catalog-content\") pod \"community-operators-6c6m2\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.630108 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7gvq\" (UniqueName: \"kubernetes.io/projected/da8155c3-0290-44f3-b1da-e95d2dcee000-kube-api-access-b7gvq\") pod \"community-operators-6c6m2\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.630176 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-utilities\") pod \"community-operators-6c6m2\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.630699 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.130682134 +0000 UTC m=+155.772702704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.631094 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-catalog-content\") pod \"community-operators-6c6m2\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.631870 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-utilities\") pod \"community-operators-6c6m2\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.648626 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rf9m5"] Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.700887 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" podStartSLOduration=135.700870154 podStartE2EDuration="2m15.700870154s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.695980932 +0000 UTC m=+155.338001492" watchObservedRunningTime="2025-12-03 00:26:09.700870154 +0000 UTC m=+155.342890714" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.712561 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7gvq\" (UniqueName: \"kubernetes.io/projected/da8155c3-0290-44f3-b1da-e95d2dcee000-kube-api-access-b7gvq\") pod \"community-operators-6c6m2\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.731540 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-t69lf" podStartSLOduration=135.73152342 podStartE2EDuration="2m15.73152342s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.72981728 +0000 UTC m=+155.371837830" watchObservedRunningTime="2025-12-03 00:26:09.73152342 +0000 UTC m=+155.373543980" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.735532 4912 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-kd5n5 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.22:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.735597 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" podUID="faa0cccb-7352-4752-8957-0d770dacaee5" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.22:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.736607 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.736667 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-utilities\") pod \"certified-operators-rf9m5\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.736695 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cmq2\" (UniqueName: \"kubernetes.io/projected/179c471f-de9c-4cd6-85ac-b8b3380401ea-kube-api-access-4cmq2\") pod \"certified-operators-rf9m5\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.736719 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-catalog-content\") pod \"certified-operators-rf9m5\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.740117 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.240098948 +0000 UTC m=+155.882119568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.762499 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.792273 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" podStartSLOduration=135.792252826 podStartE2EDuration="2m15.792252826s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:09.783855733 +0000 UTC m=+155.425876293" watchObservedRunningTime="2025-12-03 00:26:09.792252826 +0000 UTC m=+155.434273386" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.820218 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9p49z"] Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.821177 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.845606 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.845756 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cmq2\" (UniqueName: \"kubernetes.io/projected/179c471f-de9c-4cd6-85ac-b8b3380401ea-kube-api-access-4cmq2\") pod \"certified-operators-rf9m5\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.845786 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-catalog-content\") pod \"certified-operators-rf9m5\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.845828 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-catalog-content\") pod \"community-operators-9p49z\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.845847 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k87nm\" (UniqueName: \"kubernetes.io/projected/f9edbdcf-7866-4d27-a29f-5426596b00db-kube-api-access-k87nm\") pod \"community-operators-9p49z\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.845898 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-utilities\") pod \"certified-operators-rf9m5\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.845934 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-utilities\") pod \"community-operators-9p49z\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:09 crc kubenswrapper[4912]: E1203 00:26:09.846018 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.34600461 +0000 UTC m=+155.988025170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.846806 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-catalog-content\") pod \"certified-operators-rf9m5\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.847205 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-utilities\") pod \"certified-operators-rf9m5\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.850480 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" event={"ID":"3c4ce8f2-b6c1-4603-91b7-47e38cd693c2","Type":"ContainerStarted","Data":"23992d8e22df6a0a9209408673643f53f6b834954220fa88826c06ea336261ae"} Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.865661 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9p49z"] Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.891097 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vldb5" event={"ID":"82f9b623-024f-45e6-b876-f9718bc8aba6","Type":"ContainerStarted","Data":"5be768dc077dce99e900e17b1d3db326220ea5e0fed97710c86b4d03234e9aed"} Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.897918 4912 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dgskz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.897963 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.916353 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:09 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:09 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:09 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.916420 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.921193 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cmq2\" (UniqueName: \"kubernetes.io/projected/179c471f-de9c-4cd6-85ac-b8b3380401ea-kube-api-access-4cmq2\") pod \"certified-operators-rf9m5\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.947203 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-utilities\") pod \"community-operators-9p49z\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.947692 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-catalog-content\") pod \"community-operators-9p49z\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.947802 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k87nm\" (UniqueName: \"kubernetes.io/projected/f9edbdcf-7866-4d27-a29f-5426596b00db-kube-api-access-k87nm\") pod \"community-operators-9p49z\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.948149 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:09 crc kubenswrapper[4912]: I1203 00:26:09.955629 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-catalog-content\") pod \"community-operators-9p49z\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:09.996499 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.496477531 +0000 UTC m=+156.138498091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.006161 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.029420 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-utilities\") pod \"community-operators-9p49z\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.053984 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.054255 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.55422447 +0000 UTC m=+156.196245030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.054302 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.054790 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.554778286 +0000 UTC m=+156.196798846 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.158099 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.158409 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.658393222 +0000 UTC m=+156.300413782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.165606 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k87nm\" (UniqueName: \"kubernetes.io/projected/f9edbdcf-7866-4d27-a29f-5426596b00db-kube-api-access-k87nm\") pod \"community-operators-9p49z\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.189831 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.266912 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.267351 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.767333182 +0000 UTC m=+156.409353742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.376598 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.377177 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.877159448 +0000 UTC m=+156.519180008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.467934 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" podStartSLOduration=137.467917391 podStartE2EDuration="2m17.467917391s" podCreationTimestamp="2025-12-03 00:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:10.326783551 +0000 UTC m=+155.968804121" watchObservedRunningTime="2025-12-03 00:26:10.467917391 +0000 UTC m=+156.109937961" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.489503 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.489840 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:10.989828905 +0000 UTC m=+156.631849465 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.593801 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.594481 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.596154 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.596546 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.09653131 +0000 UTC m=+156.738551870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.596576 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.596855 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.096848419 +0000 UTC m=+156.738868979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.597744 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.609303 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.664025 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.701089 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.701343 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b0f63d54-1e79-4505-9642-c803d2ac0075-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b0f63d54-1e79-4505-9642-c803d2ac0075\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.701386 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b0f63d54-1e79-4505-9642-c803d2ac0075-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b0f63d54-1e79-4505-9642-c803d2ac0075\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.701539 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.201523155 +0000 UTC m=+156.843543715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.802644 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b0f63d54-1e79-4505-9642-c803d2ac0075-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b0f63d54-1e79-4505-9642-c803d2ac0075\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.802756 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b0f63d54-1e79-4505-9642-c803d2ac0075-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b0f63d54-1e79-4505-9642-c803d2ac0075\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.802787 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.803058 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.303046911 +0000 UTC m=+156.945067461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.803193 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b0f63d54-1e79-4505-9642-c803d2ac0075-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b0f63d54-1e79-4505-9642-c803d2ac0075\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.865952 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lqdfv"] Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.894549 4912 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-cns6n container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.894622 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" podUID="edb6b379-aeff-470a-a09d-540d6acc0efe" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.903440 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:10 crc kubenswrapper[4912]: E1203 00:26:10.903907 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.403887957 +0000 UTC m=+157.045908517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.921146 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:10 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:10 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:10 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:10 crc kubenswrapper[4912]: I1203 00:26:10.921235 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.009309 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.014750 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.514732161 +0000 UTC m=+157.156752721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.055935 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.056937 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.070905 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b0f63d54-1e79-4505-9642-c803d2ac0075-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b0f63d54-1e79-4505-9642-c803d2ac0075\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.111082 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.111380 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.611364985 +0000 UTC m=+157.253385545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.208024 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b5vr9" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.209927 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.209978 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.209980 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.210017 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.210366 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.210462 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.212120 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.213117 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.713101317 +0000 UTC m=+157.355121877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.214613 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.230543 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k5xn5"] Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.250208 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.261088 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.272485 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5xn5"] Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.289029 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.313814 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6c6m2"] Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.333136 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.333334 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-catalog-content\") pod \"redhat-marketplace-k5xn5\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.348569 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.848536332 +0000 UTC m=+157.490556892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.348712 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.349075 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.849066198 +0000 UTC m=+157.491086758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.349504 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-utilities\") pod \"redhat-marketplace-k5xn5\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.349548 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6vhn\" (UniqueName: \"kubernetes.io/projected/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-kube-api-access-l6vhn\") pod \"redhat-marketplace-k5xn5\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.363717 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.363779 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.384217 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.384597 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.446820 4912 patch_prober.go:28] interesting pod/console-f9d7485db-cg7ls container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.446860 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-cg7ls" podUID="b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.451332 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.451686 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-utilities\") pod \"redhat-marketplace-k5xn5\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.451718 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6vhn\" (UniqueName: \"kubernetes.io/projected/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-kube-api-access-l6vhn\") pod \"redhat-marketplace-k5xn5\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.451786 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-catalog-content\") pod \"redhat-marketplace-k5xn5\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.454567 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:11.954547698 +0000 UTC m=+157.596568258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.522401 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6vhn\" (UniqueName: \"kubernetes.io/projected/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-kube-api-access-l6vhn\") pod \"redhat-marketplace-k5xn5\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.554560 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.554934 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.05492289 +0000 UTC m=+157.696943450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.589792 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zvg59"] Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.600944 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.622501 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvg59"] Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.634170 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-catalog-content\") pod \"redhat-marketplace-k5xn5\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.635830 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-utilities\") pod \"redhat-marketplace-k5xn5\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.655594 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.655941 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-utilities\") pod \"redhat-marketplace-zvg59\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.656060 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmjq4\" (UniqueName: \"kubernetes.io/projected/616eb06f-63d7-46e2-a8b0-5162a2e159a2-kube-api-access-jmjq4\") pod \"redhat-marketplace-zvg59\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.656079 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-catalog-content\") pod \"redhat-marketplace-zvg59\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.656218 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.156200848 +0000 UTC m=+157.798221398 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.689939 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rf9m5"] Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.718863 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.758275 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmjq4\" (UniqueName: \"kubernetes.io/projected/616eb06f-63d7-46e2-a8b0-5162a2e159a2-kube-api-access-jmjq4\") pod \"redhat-marketplace-zvg59\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.758320 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-catalog-content\") pod \"redhat-marketplace-zvg59\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.758415 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-utilities\") pod \"redhat-marketplace-zvg59\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.758467 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.758796 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.258781714 +0000 UTC m=+157.900802274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.759083 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-catalog-content\") pod \"redhat-marketplace-zvg59\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.759236 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-utilities\") pod \"redhat-marketplace-zvg59\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.815547 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmjq4\" (UniqueName: \"kubernetes.io/projected/616eb06f-63d7-46e2-a8b0-5162a2e159a2-kube-api-access-jmjq4\") pod \"redhat-marketplace-zvg59\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.843772 4912 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.859934 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.860663 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.360646079 +0000 UTC m=+158.002666629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.892854 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.902161 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.912598 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:11 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:11 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:11 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.912640 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.917875 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.962907 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:11 crc kubenswrapper[4912]: E1203 00:26:11.963896 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.463881424 +0000 UTC m=+158.105902074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:11 crc kubenswrapper[4912]: I1203 00:26:11.962911 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9p49z"] Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.040465 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf9m5" event={"ID":"179c471f-de9c-4cd6-85ac-b8b3380401ea","Type":"ContainerStarted","Data":"f0a4a1486e6ac2cf6da1b6a26ee85eb9bc950027c8ff3819fadd6b51872c36a4"} Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.050536 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vldb5" event={"ID":"82f9b623-024f-45e6-b876-f9718bc8aba6","Type":"ContainerStarted","Data":"4e2fe1ce6316c9b3bc09fdf3208ce33470ed09d32cfbf4cd6c215f4a6c09e71a"} Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.068008 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:12 crc kubenswrapper[4912]: E1203 00:26:12.068294 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.568278912 +0000 UTC m=+158.210299472 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:12 crc kubenswrapper[4912]: W1203 00:26:12.079737 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9edbdcf_7866_4d27_a29f_5426596b00db.slice/crio-a3c0776f93a2e6f75237a66f274a52a03d145742516242c944f7171270f8e181 WatchSource:0}: Error finding container a3c0776f93a2e6f75237a66f274a52a03d145742516242c944f7171270f8e181: Status 404 returned error can't find the container with id a3c0776f93a2e6f75237a66f274a52a03d145742516242c944f7171270f8e181 Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.081485 4912 generic.go:334] "Generic (PLEG): container finished" podID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerID="5bbfd84cd7b31f524119d50728a4e55a185f6d46162d8c1c8e2f78d4f369ba99" exitCode=0 Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.081560 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6c6m2" event={"ID":"da8155c3-0290-44f3-b1da-e95d2dcee000","Type":"ContainerDied","Data":"5bbfd84cd7b31f524119d50728a4e55a185f6d46162d8c1c8e2f78d4f369ba99"} Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.081592 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6c6m2" event={"ID":"da8155c3-0290-44f3-b1da-e95d2dcee000","Type":"ContainerStarted","Data":"ea5f298bcce23ad702183f6170f2ae9fbc2f1ab4e74aa14e400655a6880ff0e0"} Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.093858 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.113783 4912 generic.go:334] "Generic (PLEG): container finished" podID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerID="c712ff215cb58f2d2b0e28e6f043db095cb833bf12ca25237e9b5ee8a2235fe1" exitCode=0 Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.114860 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lqdfv" event={"ID":"9e57c9b1-dd22-4cc5-ac09-f010c8030170","Type":"ContainerDied","Data":"c712ff215cb58f2d2b0e28e6f043db095cb833bf12ca25237e9b5ee8a2235fe1"} Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.114890 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lqdfv" event={"ID":"9e57c9b1-dd22-4cc5-ac09-f010c8030170","Type":"ContainerStarted","Data":"c5b02ccb545bedc42f4b6a55418471016e9e43f1ed38a7fe468632080ec3a994"} Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.139569 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-xlvsd" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.177880 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:12 crc kubenswrapper[4912]: E1203 00:26:12.178406 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.678393236 +0000 UTC m=+158.320413796 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.227565 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.281297 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:12 crc kubenswrapper[4912]: E1203 00:26:12.282823 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.782805575 +0000 UTC m=+158.424826135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.384095 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:12 crc kubenswrapper[4912]: E1203 00:26:12.384730 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.884715081 +0000 UTC m=+158.526735641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.469610 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-cns6n" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.474262 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pzmmk"] Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.487974 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:12 crc kubenswrapper[4912]: E1203 00:26:12.488291 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:12.988276616 +0000 UTC m=+158.630297176 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.490630 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.496376 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.504127 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pzmmk"] Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.595397 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.595712 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4kq4\" (UniqueName: \"kubernetes.io/projected/c13ea42e-9342-428a-8d6d-4fea05277491-kube-api-access-c4kq4\") pod \"redhat-operators-pzmmk\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.595755 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-catalog-content\") pod \"redhat-operators-pzmmk\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.595926 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-utilities\") pod \"redhat-operators-pzmmk\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: E1203 00:26:12.597960 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 00:26:13.097946927 +0000 UTC m=+158.739967487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-vbppn" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.703478 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.705906 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.706186 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-utilities\") pod \"redhat-operators-pzmmk\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.706243 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4kq4\" (UniqueName: \"kubernetes.io/projected/c13ea42e-9342-428a-8d6d-4fea05277491-kube-api-access-c4kq4\") pod \"redhat-operators-pzmmk\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.706268 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-catalog-content\") pod \"redhat-operators-pzmmk\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.709034 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-catalog-content\") pod \"redhat-operators-pzmmk\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: E1203 00:26:12.709114 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 00:26:13.20907752 +0000 UTC m=+158.851098080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.709379 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-utilities\") pod \"redhat-operators-pzmmk\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.749674 4912 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T00:26:11.843800402Z","Handler":null,"Name":""} Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.773675 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4kq4\" (UniqueName: \"kubernetes.io/projected/c13ea42e-9342-428a-8d6d-4fea05277491-kube-api-access-c4kq4\") pod \"redhat-operators-pzmmk\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.787993 4912 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.788046 4912 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.810241 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.821578 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kstcn"] Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.823003 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.840195 4912 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.840234 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.843399 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kstcn"] Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.919217 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvmqz\" (UniqueName: \"kubernetes.io/projected/668355d3-1254-4898-a4e9-425b8f480dfd-kube-api-access-bvmqz\") pod \"redhat-operators-kstcn\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.919571 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-catalog-content\") pod \"redhat-operators-kstcn\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.919655 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-utilities\") pod \"redhat-operators-kstcn\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.940610 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:12 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:12 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:12 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:12 crc kubenswrapper[4912]: I1203 00:26:12.940670 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.017699 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.018309 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-vbppn\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.027848 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvmqz\" (UniqueName: \"kubernetes.io/projected/668355d3-1254-4898-a4e9-425b8f480dfd-kube-api-access-bvmqz\") pod \"redhat-operators-kstcn\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.027900 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-catalog-content\") pod \"redhat-operators-kstcn\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.027973 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-utilities\") pod \"redhat-operators-kstcn\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.028471 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-utilities\") pod \"redhat-operators-kstcn\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.029312 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-catalog-content\") pod \"redhat-operators-kstcn\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.101003 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvmqz\" (UniqueName: \"kubernetes.io/projected/668355d3-1254-4898-a4e9-425b8f480dfd-kube-api-access-bvmqz\") pod \"redhat-operators-kstcn\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.107591 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.132172 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.168841 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5xn5"] Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.191802 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.192947 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b0f63d54-1e79-4505-9642-c803d2ac0075","Type":"ContainerStarted","Data":"5069dbc27d0d7388572b537dc09e6ee6f0b82aa4667bd51410ffe82161344346"} Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.219821 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vldb5" event={"ID":"82f9b623-024f-45e6-b876-f9718bc8aba6","Type":"ContainerStarted","Data":"d620f255578359685ce8ff867a13b7f82d612f74cb595e757ea3eecf252b3dd6"} Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.224891 4912 generic.go:334] "Generic (PLEG): container finished" podID="f6a647db-07ba-49f6-9150-0a038d4fbee8" containerID="40ce820d269d557ca9832e26779b1fc816bf349512285458c543914afb6d317b" exitCode=0 Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.224949 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" event={"ID":"f6a647db-07ba-49f6-9150-0a038d4fbee8","Type":"ContainerDied","Data":"40ce820d269d557ca9832e26779b1fc816bf349512285458c543914afb6d317b"} Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.242173 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.257012 4912 generic.go:334] "Generic (PLEG): container finished" podID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerID="1d79119955b934defb3407c59f649b7ae3b9559a476a96e4428fd842dc936cc6" exitCode=0 Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.257096 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p49z" event={"ID":"f9edbdcf-7866-4d27-a29f-5426596b00db","Type":"ContainerDied","Data":"1d79119955b934defb3407c59f649b7ae3b9559a476a96e4428fd842dc936cc6"} Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.257122 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p49z" event={"ID":"f9edbdcf-7866-4d27-a29f-5426596b00db","Type":"ContainerStarted","Data":"a3c0776f93a2e6f75237a66f274a52a03d145742516242c944f7171270f8e181"} Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.270796 4912 generic.go:334] "Generic (PLEG): container finished" podID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerID="26678c1c72f5e46cf74cba167b00d69911bf529e770d947d9ea2343c980c22fb" exitCode=0 Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.271737 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf9m5" event={"ID":"179c471f-de9c-4cd6-85ac-b8b3380401ea","Type":"ContainerDied","Data":"26678c1c72f5e46cf74cba167b00d69911bf529e770d947d9ea2343c980c22fb"} Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.272368 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-vldb5" podStartSLOduration=14.272356706 podStartE2EDuration="14.272356706s" podCreationTimestamp="2025-12-03 00:25:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:13.269570135 +0000 UTC m=+158.911590695" watchObservedRunningTime="2025-12-03 00:26:13.272356706 +0000 UTC m=+158.914377266" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.289551 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-c65h9" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.384454 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvg59"] Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.839374 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbppn"] Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.850619 4912 patch_prober.go:28] interesting pod/apiserver-76f77b778f-lg5fq container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]log ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]etcd ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/max-in-flight-filter ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 00:26:13 crc kubenswrapper[4912]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 00:26:13 crc kubenswrapper[4912]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/openshift.io-startinformers ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 00:26:13 crc kubenswrapper[4912]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 00:26:13 crc kubenswrapper[4912]: livez check failed Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.850909 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" podUID="3c4ce8f2-b6c1-4603-91b7-47e38cd693c2" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.880594 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pzmmk"] Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.911520 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:13 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:13 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:13 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:13 crc kubenswrapper[4912]: I1203 00:26:13.911607 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.201558 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kstcn"] Dec 03 00:26:14 crc kubenswrapper[4912]: W1203 00:26:14.228017 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod668355d3_1254_4898_a4e9_425b8f480dfd.slice/crio-7ef0721aaa32c2c5ec6ed2dbca3d6294a6854641ee53994cc2ea7162b9652064 WatchSource:0}: Error finding container 7ef0721aaa32c2c5ec6ed2dbca3d6294a6854641ee53994cc2ea7162b9652064: Status 404 returned error can't find the container with id 7ef0721aaa32c2c5ec6ed2dbca3d6294a6854641ee53994cc2ea7162b9652064 Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.296064 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" event={"ID":"a64deb49-4066-4d06-92fb-cd522ed7e496","Type":"ContainerStarted","Data":"1617c385f580def901687574248c144ec5d107f4121095eebd10d7c5c5f3f184"} Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.299643 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kstcn" event={"ID":"668355d3-1254-4898-a4e9-425b8f480dfd","Type":"ContainerStarted","Data":"7ef0721aaa32c2c5ec6ed2dbca3d6294a6854641ee53994cc2ea7162b9652064"} Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.302462 4912 generic.go:334] "Generic (PLEG): container finished" podID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerID="f5849234f6357e0d46b42e30436d7074f9c52e6b126359331636f39d44c266c0" exitCode=0 Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.302522 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvg59" event={"ID":"616eb06f-63d7-46e2-a8b0-5162a2e159a2","Type":"ContainerDied","Data":"f5849234f6357e0d46b42e30436d7074f9c52e6b126359331636f39d44c266c0"} Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.302548 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvg59" event={"ID":"616eb06f-63d7-46e2-a8b0-5162a2e159a2","Type":"ContainerStarted","Data":"0d9771f13a7c78bc19f6f348990bd0ee1125a92973038b84e73fb3050b06454b"} Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.310244 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzmmk" event={"ID":"c13ea42e-9342-428a-8d6d-4fea05277491","Type":"ContainerStarted","Data":"aa927155ec38e791f42077e456b8cdc07b7d19b4e6b819af4e801d1aa8af01de"} Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.314762 4912 generic.go:334] "Generic (PLEG): container finished" podID="b0f63d54-1e79-4505-9642-c803d2ac0075" containerID="8d7f6959afea664fc4507b369b8f8aabc7aa1be151cb7ed2b5be5e15632fa3bb" exitCode=0 Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.314862 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b0f63d54-1e79-4505-9642-c803d2ac0075","Type":"ContainerDied","Data":"8d7f6959afea664fc4507b369b8f8aabc7aa1be151cb7ed2b5be5e15632fa3bb"} Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.326780 4912 generic.go:334] "Generic (PLEG): container finished" podID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerID="1fc0ea09da6279922d43a9a483e82a2d947634813381f7aaf1c25aced86081d6" exitCode=0 Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.326847 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5xn5" event={"ID":"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01","Type":"ContainerDied","Data":"1fc0ea09da6279922d43a9a483e82a2d947634813381f7aaf1c25aced86081d6"} Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.326938 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5xn5" event={"ID":"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01","Type":"ContainerStarted","Data":"2893a29b6e768df7495f4bdf462540c54e545ae2e36465093946dd778202731b"} Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.641134 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.917704 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:14 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:14 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:14 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.917770 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:14 crc kubenswrapper[4912]: I1203 00:26:14.935711 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.086757 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjbml\" (UniqueName: \"kubernetes.io/projected/f6a647db-07ba-49f6-9150-0a038d4fbee8-kube-api-access-mjbml\") pod \"f6a647db-07ba-49f6-9150-0a038d4fbee8\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.086811 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6a647db-07ba-49f6-9150-0a038d4fbee8-config-volume\") pod \"f6a647db-07ba-49f6-9150-0a038d4fbee8\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.086900 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6a647db-07ba-49f6-9150-0a038d4fbee8-secret-volume\") pod \"f6a647db-07ba-49f6-9150-0a038d4fbee8\" (UID: \"f6a647db-07ba-49f6-9150-0a038d4fbee8\") " Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.087680 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6a647db-07ba-49f6-9150-0a038d4fbee8-config-volume" (OuterVolumeSpecName: "config-volume") pod "f6a647db-07ba-49f6-9150-0a038d4fbee8" (UID: "f6a647db-07ba-49f6-9150-0a038d4fbee8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.128258 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6a647db-07ba-49f6-9150-0a038d4fbee8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f6a647db-07ba-49f6-9150-0a038d4fbee8" (UID: "f6a647db-07ba-49f6-9150-0a038d4fbee8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.141517 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6a647db-07ba-49f6-9150-0a038d4fbee8-kube-api-access-mjbml" (OuterVolumeSpecName: "kube-api-access-mjbml") pod "f6a647db-07ba-49f6-9150-0a038d4fbee8" (UID: "f6a647db-07ba-49f6-9150-0a038d4fbee8"). InnerVolumeSpecName "kube-api-access-mjbml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.192060 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f6a647db-07ba-49f6-9150-0a038d4fbee8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.192099 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjbml\" (UniqueName: \"kubernetes.io/projected/f6a647db-07ba-49f6-9150-0a038d4fbee8-kube-api-access-mjbml\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.192117 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f6a647db-07ba-49f6-9150-0a038d4fbee8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.376569 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.376630 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4" event={"ID":"f6a647db-07ba-49f6-9150-0a038d4fbee8","Type":"ContainerDied","Data":"5bb77a65814e7b8ebcb4549e641f2eeaf19131e6ab981d5e6fcdd4008699af09"} Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.376670 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bb77a65814e7b8ebcb4549e641f2eeaf19131e6ab981d5e6fcdd4008699af09" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.380737 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" event={"ID":"a64deb49-4066-4d06-92fb-cd522ed7e496","Type":"ContainerStarted","Data":"fa7de72ef46cf4c0312c41ac3ef480da5028328b8bd677e2f50aa4931a15e6e1"} Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.381336 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.392650 4912 generic.go:334] "Generic (PLEG): container finished" podID="668355d3-1254-4898-a4e9-425b8f480dfd" containerID="036905eca3af59e5936a201fb04c7100d52f24b312eee0778466c87170c8b666" exitCode=0 Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.392746 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kstcn" event={"ID":"668355d3-1254-4898-a4e9-425b8f480dfd","Type":"ContainerDied","Data":"036905eca3af59e5936a201fb04c7100d52f24b312eee0778466c87170c8b666"} Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.399189 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzmmk" event={"ID":"c13ea42e-9342-428a-8d6d-4fea05277491","Type":"ContainerDied","Data":"b3cbd098a44ed8f847624dd7f5446c941b2cbd63359696f25e19b016152a5085"} Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.399817 4912 generic.go:334] "Generic (PLEG): container finished" podID="c13ea42e-9342-428a-8d6d-4fea05277491" containerID="b3cbd098a44ed8f847624dd7f5446c941b2cbd63359696f25e19b016152a5085" exitCode=0 Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.421256 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" podStartSLOduration=141.421226706 podStartE2EDuration="2m21.421226706s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:15.402491255 +0000 UTC m=+161.044511825" watchObservedRunningTime="2025-12-03 00:26:15.421226706 +0000 UTC m=+161.063247266" Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.903844 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:15 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:15 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:15 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:15 crc kubenswrapper[4912]: I1203 00:26:15.903905 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.024038 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.045512 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b0f63d54-1e79-4505-9642-c803d2ac0075-kubelet-dir\") pod \"b0f63d54-1e79-4505-9642-c803d2ac0075\" (UID: \"b0f63d54-1e79-4505-9642-c803d2ac0075\") " Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.045675 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b0f63d54-1e79-4505-9642-c803d2ac0075-kube-api-access\") pod \"b0f63d54-1e79-4505-9642-c803d2ac0075\" (UID: \"b0f63d54-1e79-4505-9642-c803d2ac0075\") " Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.046763 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0f63d54-1e79-4505-9642-c803d2ac0075-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b0f63d54-1e79-4505-9642-c803d2ac0075" (UID: "b0f63d54-1e79-4505-9642-c803d2ac0075"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.058585 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f63d54-1e79-4505-9642-c803d2ac0075-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b0f63d54-1e79-4505-9642-c803d2ac0075" (UID: "b0f63d54-1e79-4505-9642-c803d2ac0075"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.146277 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b0f63d54-1e79-4505-9642-c803d2ac0075-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.146308 4912 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b0f63d54-1e79-4505-9642-c803d2ac0075-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.369979 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.378734 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-lg5fq" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.407033 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b0f63d54-1e79-4505-9642-c803d2ac0075","Type":"ContainerDied","Data":"5069dbc27d0d7388572b537dc09e6ee6f0b82aa4667bd51410ffe82161344346"} Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.407102 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5069dbc27d0d7388572b537dc09e6ee6f0b82aa4667bd51410ffe82161344346" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.407290 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.917953 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:16 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:16 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:16 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:16 crc kubenswrapper[4912]: I1203 00:26:16.918037 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.411934 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.432977 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e49d5e0a-f59c-4dc1-9140-4b255576cd72-metrics-certs\") pod \"network-metrics-daemon-kdznz\" (UID: \"e49d5e0a-f59c-4dc1-9140-4b255576cd72\") " pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.601709 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kdznz" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.736346 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 00:26:17 crc kubenswrapper[4912]: E1203 00:26:17.736677 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6a647db-07ba-49f6-9150-0a038d4fbee8" containerName="collect-profiles" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.736701 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6a647db-07ba-49f6-9150-0a038d4fbee8" containerName="collect-profiles" Dec 03 00:26:17 crc kubenswrapper[4912]: E1203 00:26:17.736748 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f63d54-1e79-4505-9642-c803d2ac0075" containerName="pruner" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.736757 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f63d54-1e79-4505-9642-c803d2ac0075" containerName="pruner" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.736894 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0f63d54-1e79-4505-9642-c803d2ac0075" containerName="pruner" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.736923 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6a647db-07ba-49f6-9150-0a038d4fbee8" containerName="collect-profiles" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.738513 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.748646 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.749082 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.770134 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.777405 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vdwv6" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.838222 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75cefecf-ae30-4214-b137-48d6846f1e08-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"75cefecf-ae30-4214-b137-48d6846f1e08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.838698 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75cefecf-ae30-4214-b137-48d6846f1e08-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"75cefecf-ae30-4214-b137-48d6846f1e08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.906051 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:17 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:17 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:17 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.906122 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.939995 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75cefecf-ae30-4214-b137-48d6846f1e08-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"75cefecf-ae30-4214-b137-48d6846f1e08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.940054 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75cefecf-ae30-4214-b137-48d6846f1e08-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"75cefecf-ae30-4214-b137-48d6846f1e08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.940167 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75cefecf-ae30-4214-b137-48d6846f1e08-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"75cefecf-ae30-4214-b137-48d6846f1e08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:17 crc kubenswrapper[4912]: I1203 00:26:17.981035 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75cefecf-ae30-4214-b137-48d6846f1e08-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"75cefecf-ae30-4214-b137-48d6846f1e08\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:18 crc kubenswrapper[4912]: I1203 00:26:18.077563 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:26:18 crc kubenswrapper[4912]: I1203 00:26:18.077626 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:26:18 crc kubenswrapper[4912]: I1203 00:26:18.103787 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:18 crc kubenswrapper[4912]: I1203 00:26:18.298488 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kdznz"] Dec 03 00:26:18 crc kubenswrapper[4912]: W1203 00:26:18.319531 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode49d5e0a_f59c_4dc1_9140_4b255576cd72.slice/crio-b4945595a3413ede66227a618c886aa7f5a869dff73544be9b90c03c18862c1b WatchSource:0}: Error finding container b4945595a3413ede66227a618c886aa7f5a869dff73544be9b90c03c18862c1b: Status 404 returned error can't find the container with id b4945595a3413ede66227a618c886aa7f5a869dff73544be9b90c03c18862c1b Dec 03 00:26:18 crc kubenswrapper[4912]: I1203 00:26:18.518132 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kdznz" event={"ID":"e49d5e0a-f59c-4dc1-9140-4b255576cd72","Type":"ContainerStarted","Data":"b4945595a3413ede66227a618c886aa7f5a869dff73544be9b90c03c18862c1b"} Dec 03 00:26:18 crc kubenswrapper[4912]: I1203 00:26:18.651868 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 00:26:18 crc kubenswrapper[4912]: I1203 00:26:18.905381 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:18 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:18 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:18 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:18 crc kubenswrapper[4912]: I1203 00:26:18.905454 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:19 crc kubenswrapper[4912]: I1203 00:26:19.531230 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"75cefecf-ae30-4214-b137-48d6846f1e08","Type":"ContainerStarted","Data":"3f6b0acf5f25379aed6fc9e08c0b4a2c3fea73c4b109135fa7307092bcefe8e6"} Dec 03 00:26:19 crc kubenswrapper[4912]: I1203 00:26:19.905018 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:19 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:19 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:19 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:19 crc kubenswrapper[4912]: I1203 00:26:19.905105 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:20 crc kubenswrapper[4912]: I1203 00:26:20.905334 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:20 crc kubenswrapper[4912]: [-]has-synced failed: reason withheld Dec 03 00:26:20 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:20 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:20 crc kubenswrapper[4912]: I1203 00:26:20.905412 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.208257 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.208640 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.208272 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.208767 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.379948 4912 patch_prober.go:28] interesting pod/console-f9d7485db-cg7ls container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.380066 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-cg7ls" podUID="b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" containerName="console" probeResult="failure" output="Get \"https://10.217.0.19:8443/health\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.554305 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"75cefecf-ae30-4214-b137-48d6846f1e08","Type":"ContainerStarted","Data":"47767ea39db4c6dbcf952af06140199800c9cbbd1e91beb0a8f0987003845855"} Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.560556 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kdznz" event={"ID":"e49d5e0a-f59c-4dc1-9140-4b255576cd72","Type":"ContainerStarted","Data":"fd1a7fa84dc8632fe3ebde8eea9e8ee3cb865a4a2cbf1f3f2f9a1f86b2252a83"} Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.579590 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.5795687019999995 podStartE2EDuration="4.579568702s" podCreationTimestamp="2025-12-03 00:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:21.576298658 +0000 UTC m=+167.218319218" watchObservedRunningTime="2025-12-03 00:26:21.579568702 +0000 UTC m=+167.221589262" Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.905168 4912 patch_prober.go:28] interesting pod/router-default-5444994796-dddxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 00:26:21 crc kubenswrapper[4912]: [+]has-synced ok Dec 03 00:26:21 crc kubenswrapper[4912]: [+]process-running ok Dec 03 00:26:21 crc kubenswrapper[4912]: healthz check failed Dec 03 00:26:21 crc kubenswrapper[4912]: I1203 00:26:21.905321 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-dddxf" podUID="8a39ab4f-ca7e-42ef-afe5-43ab6076dcb4" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 00:26:22 crc kubenswrapper[4912]: I1203 00:26:22.570785 4912 generic.go:334] "Generic (PLEG): container finished" podID="75cefecf-ae30-4214-b137-48d6846f1e08" containerID="47767ea39db4c6dbcf952af06140199800c9cbbd1e91beb0a8f0987003845855" exitCode=0 Dec 03 00:26:22 crc kubenswrapper[4912]: I1203 00:26:22.600666 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"75cefecf-ae30-4214-b137-48d6846f1e08","Type":"ContainerDied","Data":"47767ea39db4c6dbcf952af06140199800c9cbbd1e91beb0a8f0987003845855"} Dec 03 00:26:22 crc kubenswrapper[4912]: I1203 00:26:22.908451 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:26:22 crc kubenswrapper[4912]: I1203 00:26:22.921587 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-dddxf" Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.208232 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.208853 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.208455 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.208902 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-b5vr9" Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.208948 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.209316 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.209358 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.209574 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"0e2f9a41224c4e26226217fa58b4a1c0d8dc10266e6cb15fe5b07b2309076ecc"} pod="openshift-console/downloads-7954f5f757-b5vr9" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.209649 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" containerID="cri-o://0e2f9a41224c4e26226217fa58b4a1c0d8dc10266e6cb15fe5b07b2309076ecc" gracePeriod=2 Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.384279 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:26:31 crc kubenswrapper[4912]: I1203 00:26:31.387878 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.403209 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.509328 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75cefecf-ae30-4214-b137-48d6846f1e08-kubelet-dir\") pod \"75cefecf-ae30-4214-b137-48d6846f1e08\" (UID: \"75cefecf-ae30-4214-b137-48d6846f1e08\") " Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.509477 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75cefecf-ae30-4214-b137-48d6846f1e08-kube-api-access\") pod \"75cefecf-ae30-4214-b137-48d6846f1e08\" (UID: \"75cefecf-ae30-4214-b137-48d6846f1e08\") " Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.510836 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75cefecf-ae30-4214-b137-48d6846f1e08-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "75cefecf-ae30-4214-b137-48d6846f1e08" (UID: "75cefecf-ae30-4214-b137-48d6846f1e08"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.530112 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75cefecf-ae30-4214-b137-48d6846f1e08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "75cefecf-ae30-4214-b137-48d6846f1e08" (UID: "75cefecf-ae30-4214-b137-48d6846f1e08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.611957 4912 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75cefecf-ae30-4214-b137-48d6846f1e08-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.612280 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75cefecf-ae30-4214-b137-48d6846f1e08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.737534 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.737499 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"75cefecf-ae30-4214-b137-48d6846f1e08","Type":"ContainerDied","Data":"3f6b0acf5f25379aed6fc9e08c0b4a2c3fea73c4b109135fa7307092bcefe8e6"} Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.737666 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f6b0acf5f25379aed6fc9e08c0b4a2c3fea73c4b109135fa7307092bcefe8e6" Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.745003 4912 generic.go:334] "Generic (PLEG): container finished" podID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerID="0e2f9a41224c4e26226217fa58b4a1c0d8dc10266e6cb15fe5b07b2309076ecc" exitCode=0 Dec 03 00:26:32 crc kubenswrapper[4912]: I1203 00:26:32.745045 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b5vr9" event={"ID":"61f377a2-06fc-427d-9c3e-38f9b3473e23","Type":"ContainerDied","Data":"0e2f9a41224c4e26226217fa58b4a1c0d8dc10266e6cb15fe5b07b2309076ecc"} Dec 03 00:26:33 crc kubenswrapper[4912]: I1203 00:26:33.117560 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:26:35 crc kubenswrapper[4912]: I1203 00:26:35.768616 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kdznz" event={"ID":"e49d5e0a-f59c-4dc1-9140-4b255576cd72","Type":"ContainerStarted","Data":"41aef295473b6de655eb2eb490caf399f27af5ff1702f1e30cd4396040c79745"} Dec 03 00:26:41 crc kubenswrapper[4912]: I1203 00:26:41.070621 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 00:26:41 crc kubenswrapper[4912]: I1203 00:26:41.090287 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-kdznz" podStartSLOduration=167.090241553 podStartE2EDuration="2m47.090241553s" podCreationTimestamp="2025-12-03 00:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:26:35.786330502 +0000 UTC m=+181.428351072" watchObservedRunningTime="2025-12-03 00:26:41.090241553 +0000 UTC m=+186.732262133" Dec 03 00:26:41 crc kubenswrapper[4912]: I1203 00:26:41.208689 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:41 crc kubenswrapper[4912]: I1203 00:26:41.208757 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:41 crc kubenswrapper[4912]: I1203 00:26:41.905656 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wdb6p" Dec 03 00:26:42 crc kubenswrapper[4912]: I1203 00:26:42.809953 4912 generic.go:334] "Generic (PLEG): container finished" podID="2839435a-a464-473c-94ad-7673ad00630b" containerID="1a987a1607172aa64b660afd6ce811b1aefc7dbbd59a4da36e6868ff179de61a" exitCode=0 Dec 03 00:26:42 crc kubenswrapper[4912]: I1203 00:26:42.810000 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29412000-5584k" event={"ID":"2839435a-a464-473c-94ad-7673ad00630b","Type":"ContainerDied","Data":"1a987a1607172aa64b660afd6ce811b1aefc7dbbd59a4da36e6868ff179de61a"} Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.679780 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.811693 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt7ft\" (UniqueName: \"kubernetes.io/projected/2839435a-a464-473c-94ad-7673ad00630b-kube-api-access-zt7ft\") pod \"2839435a-a464-473c-94ad-7673ad00630b\" (UID: \"2839435a-a464-473c-94ad-7673ad00630b\") " Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.811981 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2839435a-a464-473c-94ad-7673ad00630b-serviceca\") pod \"2839435a-a464-473c-94ad-7673ad00630b\" (UID: \"2839435a-a464-473c-94ad-7673ad00630b\") " Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.812625 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2839435a-a464-473c-94ad-7673ad00630b-serviceca" (OuterVolumeSpecName: "serviceca") pod "2839435a-a464-473c-94ad-7673ad00630b" (UID: "2839435a-a464-473c-94ad-7673ad00630b"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.818871 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2839435a-a464-473c-94ad-7673ad00630b-kube-api-access-zt7ft" (OuterVolumeSpecName: "kube-api-access-zt7ft") pod "2839435a-a464-473c-94ad-7673ad00630b" (UID: "2839435a-a464-473c-94ad-7673ad00630b"). InnerVolumeSpecName "kube-api-access-zt7ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.829780 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29412000-5584k" event={"ID":"2839435a-a464-473c-94ad-7673ad00630b","Type":"ContainerDied","Data":"61b58ffc85505a8bb13404bdeba67fd7db900bdd1edad23f8d68de78f3430e86"} Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.829816 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61b58ffc85505a8bb13404bdeba67fd7db900bdd1edad23f8d68de78f3430e86" Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.830034 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29412000-5584k" Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.914882 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt7ft\" (UniqueName: \"kubernetes.io/projected/2839435a-a464-473c-94ad-7673ad00630b-kube-api-access-zt7ft\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:46 crc kubenswrapper[4912]: I1203 00:26:46.914916 4912 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2839435a-a464-473c-94ad-7673ad00630b-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 00:26:48 crc kubenswrapper[4912]: I1203 00:26:48.078008 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:26:48 crc kubenswrapper[4912]: I1203 00:26:48.078078 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:26:51 crc kubenswrapper[4912]: I1203 00:26:51.209362 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:26:51 crc kubenswrapper[4912]: I1203 00:26:51.209729 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:26:52 crc kubenswrapper[4912]: E1203 00:26:52.157377 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 00:26:52 crc kubenswrapper[4912]: E1203 00:26:52.157829 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4szh6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-lqdfv_openshift-marketplace(9e57c9b1-dd22-4cc5-ac09-f010c8030170): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:26:52 crc kubenswrapper[4912]: E1203 00:26:52.159053 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-lqdfv" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" Dec 03 00:26:53 crc kubenswrapper[4912]: E1203 00:26:53.144973 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 00:26:53 crc kubenswrapper[4912]: E1203 00:26:53.145206 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b7gvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6c6m2_openshift-marketplace(da8155c3-0290-44f3-b1da-e95d2dcee000): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:26:53 crc kubenswrapper[4912]: E1203 00:26:53.146386 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6c6m2" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.717962 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 00:26:53 crc kubenswrapper[4912]: E1203 00:26:53.718395 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75cefecf-ae30-4214-b137-48d6846f1e08" containerName="pruner" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.718448 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="75cefecf-ae30-4214-b137-48d6846f1e08" containerName="pruner" Dec 03 00:26:53 crc kubenswrapper[4912]: E1203 00:26:53.718476 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2839435a-a464-473c-94ad-7673ad00630b" containerName="image-pruner" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.718486 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2839435a-a464-473c-94ad-7673ad00630b" containerName="image-pruner" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.718650 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="75cefecf-ae30-4214-b137-48d6846f1e08" containerName="pruner" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.718669 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2839435a-a464-473c-94ad-7673ad00630b" containerName="image-pruner" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.719337 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.725060 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.725818 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.732044 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.821832 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83ecbc09-cfac-405a-8385-8e2343776d29-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"83ecbc09-cfac-405a-8385-8e2343776d29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.821880 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83ecbc09-cfac-405a-8385-8e2343776d29-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"83ecbc09-cfac-405a-8385-8e2343776d29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.923629 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83ecbc09-cfac-405a-8385-8e2343776d29-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"83ecbc09-cfac-405a-8385-8e2343776d29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.923724 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83ecbc09-cfac-405a-8385-8e2343776d29-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"83ecbc09-cfac-405a-8385-8e2343776d29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.923781 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83ecbc09-cfac-405a-8385-8e2343776d29-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"83ecbc09-cfac-405a-8385-8e2343776d29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:26:53 crc kubenswrapper[4912]: I1203 00:26:53.946055 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83ecbc09-cfac-405a-8385-8e2343776d29-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"83ecbc09-cfac-405a-8385-8e2343776d29\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:26:54 crc kubenswrapper[4912]: I1203 00:26:54.055178 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:26:57 crc kubenswrapper[4912]: E1203 00:26:57.651054 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-lqdfv" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" Dec 03 00:26:57 crc kubenswrapper[4912]: E1203 00:26:57.651137 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6c6m2" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" Dec 03 00:26:57 crc kubenswrapper[4912]: E1203 00:26:57.673726 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 00:26:57 crc kubenswrapper[4912]: E1203 00:26:57.673940 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bvmqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kstcn_openshift-marketplace(668355d3-1254-4898-a4e9-425b8f480dfd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:26:57 crc kubenswrapper[4912]: E1203 00:26:57.675226 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kstcn" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" Dec 03 00:26:57 crc kubenswrapper[4912]: E1203 00:26:57.795857 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 00:26:57 crc kubenswrapper[4912]: E1203 00:26:57.796013 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k87nm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9p49z_openshift-marketplace(f9edbdcf-7866-4d27-a29f-5426596b00db): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:26:57 crc kubenswrapper[4912]: E1203 00:26:57.797196 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9p49z" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.521325 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.523163 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.531417 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.590892 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-var-lock\") pod \"installer-9-crc\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.590969 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-kubelet-dir\") pod \"installer-9-crc\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.590990 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07a517b0-e70e-4eea-ba9f-3f2b18153008-kube-api-access\") pod \"installer-9-crc\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.692039 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-var-lock\") pod \"installer-9-crc\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.692136 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-kubelet-dir\") pod \"installer-9-crc\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.692155 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07a517b0-e70e-4eea-ba9f-3f2b18153008-kube-api-access\") pod \"installer-9-crc\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.692541 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-var-lock\") pod \"installer-9-crc\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.693875 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-kubelet-dir\") pod \"installer-9-crc\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.710208 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07a517b0-e70e-4eea-ba9f-3f2b18153008-kube-api-access\") pod \"installer-9-crc\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:26:58 crc kubenswrapper[4912]: I1203 00:26:58.854092 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.112206 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kstcn" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.112238 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9p49z" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.208989 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.209228 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.444238 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.444401 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jmjq4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zvg59_openshift-marketplace(616eb06f-63d7-46e2-a8b0-5162a2e159a2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.445669 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-zvg59" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.522558 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.522731 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l6vhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-k5xn5_openshift-marketplace(0c0811e2-2b5f-4dd9-9b2f-6039a3297e01): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.524071 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-k5xn5" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.532157 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 00:27:01 crc kubenswrapper[4912]: W1203 00:27:01.538372 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod83ecbc09_cfac_405a_8385_8e2343776d29.slice/crio-6880a695a1c9bdd15948e23ab16c4d59cc6d5d54d3e37e1abbcb3dec6b10b531 WatchSource:0}: Error finding container 6880a695a1c9bdd15948e23ab16c4d59cc6d5d54d3e37e1abbcb3dec6b10b531: Status 404 returned error can't find the container with id 6880a695a1c9bdd15948e23ab16c4d59cc6d5d54d3e37e1abbcb3dec6b10b531 Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.596531 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 00:27:01 crc kubenswrapper[4912]: W1203 00:27:01.602050 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod07a517b0_e70e_4eea_ba9f_3f2b18153008.slice/crio-77517e41cef4bca79468f35377b70f28e519ab681110025418f841467a32c844 WatchSource:0}: Error finding container 77517e41cef4bca79468f35377b70f28e519ab681110025418f841467a32c844: Status 404 returned error can't find the container with id 77517e41cef4bca79468f35377b70f28e519ab681110025418f841467a32c844 Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.911336 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b5vr9" event={"ID":"61f377a2-06fc-427d-9c3e-38f9b3473e23","Type":"ContainerStarted","Data":"9e9c1b490946f2c230bb167696f75dd406a89c8806407850e5d8d800ad45048a"} Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.911658 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b5vr9" Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.911956 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.911996 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.912912 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"83ecbc09-cfac-405a-8385-8e2343776d29","Type":"ContainerStarted","Data":"6880a695a1c9bdd15948e23ab16c4d59cc6d5d54d3e37e1abbcb3dec6b10b531"} Dec 03 00:27:01 crc kubenswrapper[4912]: I1203 00:27:01.915519 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"07a517b0-e70e-4eea-ba9f-3f2b18153008","Type":"ContainerStarted","Data":"77517e41cef4bca79468f35377b70f28e519ab681110025418f841467a32c844"} Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.916632 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zvg59" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" Dec 03 00:27:01 crc kubenswrapper[4912]: E1203 00:27:01.917119 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-k5xn5" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" Dec 03 00:27:02 crc kubenswrapper[4912]: E1203 00:27:02.111959 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 00:27:02 crc kubenswrapper[4912]: E1203 00:27:02.112172 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4cmq2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rf9m5_openshift-marketplace(179c471f-de9c-4cd6-85ac-b8b3380401ea): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:27:02 crc kubenswrapper[4912]: E1203 00:27:02.113384 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rf9m5" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" Dec 03 00:27:02 crc kubenswrapper[4912]: I1203 00:27:02.922557 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"83ecbc09-cfac-405a-8385-8e2343776d29","Type":"ContainerStarted","Data":"6a6580b96a83d67698e482c909fa28a6ba1ff8bf5fcfc877abd2e388d073711a"} Dec 03 00:27:02 crc kubenswrapper[4912]: I1203 00:27:02.924291 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"07a517b0-e70e-4eea-ba9f-3f2b18153008","Type":"ContainerStarted","Data":"7f5a733fb2a6a806700ceee260c0f8b7df28c2735a4609118cff55e5e0e99d32"} Dec 03 00:27:02 crc kubenswrapper[4912]: I1203 00:27:02.925100 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:27:02 crc kubenswrapper[4912]: I1203 00:27:02.925136 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:27:02 crc kubenswrapper[4912]: I1203 00:27:02.936553 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=9.936531793 podStartE2EDuration="9.936531793s" podCreationTimestamp="2025-12-03 00:26:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:27:02.934840344 +0000 UTC m=+208.576860914" watchObservedRunningTime="2025-12-03 00:27:02.936531793 +0000 UTC m=+208.578552353" Dec 03 00:27:02 crc kubenswrapper[4912]: I1203 00:27:02.971043 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=4.971020926 podStartE2EDuration="4.971020926s" podCreationTimestamp="2025-12-03 00:26:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:27:02.969179962 +0000 UTC m=+208.611200532" watchObservedRunningTime="2025-12-03 00:27:02.971020926 +0000 UTC m=+208.613041486" Dec 03 00:27:03 crc kubenswrapper[4912]: E1203 00:27:03.541073 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rf9m5" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" Dec 03 00:27:03 crc kubenswrapper[4912]: I1203 00:27:03.931636 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzmmk" event={"ID":"c13ea42e-9342-428a-8d6d-4fea05277491","Type":"ContainerStarted","Data":"530ef7e8325f79cbe06558aae67e3454e4cdb54ba98cdddf5542475d922c64a7"} Dec 03 00:27:03 crc kubenswrapper[4912]: I1203 00:27:03.933201 4912 generic.go:334] "Generic (PLEG): container finished" podID="83ecbc09-cfac-405a-8385-8e2343776d29" containerID="6a6580b96a83d67698e482c909fa28a6ba1ff8bf5fcfc877abd2e388d073711a" exitCode=0 Dec 03 00:27:03 crc kubenswrapper[4912]: I1203 00:27:03.933299 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"83ecbc09-cfac-405a-8385-8e2343776d29","Type":"ContainerDied","Data":"6a6580b96a83d67698e482c909fa28a6ba1ff8bf5fcfc877abd2e388d073711a"} Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.286094 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.378072 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83ecbc09-cfac-405a-8385-8e2343776d29-kubelet-dir\") pod \"83ecbc09-cfac-405a-8385-8e2343776d29\" (UID: \"83ecbc09-cfac-405a-8385-8e2343776d29\") " Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.378208 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/83ecbc09-cfac-405a-8385-8e2343776d29-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "83ecbc09-cfac-405a-8385-8e2343776d29" (UID: "83ecbc09-cfac-405a-8385-8e2343776d29"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.378242 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83ecbc09-cfac-405a-8385-8e2343776d29-kube-api-access\") pod \"83ecbc09-cfac-405a-8385-8e2343776d29\" (UID: \"83ecbc09-cfac-405a-8385-8e2343776d29\") " Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.378474 4912 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83ecbc09-cfac-405a-8385-8e2343776d29-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.383649 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83ecbc09-cfac-405a-8385-8e2343776d29-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "83ecbc09-cfac-405a-8385-8e2343776d29" (UID: "83ecbc09-cfac-405a-8385-8e2343776d29"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.479812 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/83ecbc09-cfac-405a-8385-8e2343776d29-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.946294 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"83ecbc09-cfac-405a-8385-8e2343776d29","Type":"ContainerDied","Data":"6880a695a1c9bdd15948e23ab16c4d59cc6d5d54d3e37e1abbcb3dec6b10b531"} Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.946339 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6880a695a1c9bdd15948e23ab16c4d59cc6d5d54d3e37e1abbcb3dec6b10b531" Dec 03 00:27:05 crc kubenswrapper[4912]: I1203 00:27:05.946458 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 00:27:07 crc kubenswrapper[4912]: I1203 00:27:07.958486 4912 generic.go:334] "Generic (PLEG): container finished" podID="c13ea42e-9342-428a-8d6d-4fea05277491" containerID="530ef7e8325f79cbe06558aae67e3454e4cdb54ba98cdddf5542475d922c64a7" exitCode=0 Dec 03 00:27:07 crc kubenswrapper[4912]: I1203 00:27:07.958569 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzmmk" event={"ID":"c13ea42e-9342-428a-8d6d-4fea05277491","Type":"ContainerDied","Data":"530ef7e8325f79cbe06558aae67e3454e4cdb54ba98cdddf5542475d922c64a7"} Dec 03 00:27:11 crc kubenswrapper[4912]: I1203 00:27:11.209816 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:27:11 crc kubenswrapper[4912]: I1203 00:27:11.210399 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:27:11 crc kubenswrapper[4912]: I1203 00:27:11.210126 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:27:11 crc kubenswrapper[4912]: I1203 00:27:11.210523 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:27:15 crc kubenswrapper[4912]: I1203 00:27:15.008285 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzmmk" event={"ID":"c13ea42e-9342-428a-8d6d-4fea05277491","Type":"ContainerStarted","Data":"cc88dacdc2b7e927cfd034a9b956d8d4fd9f64a4e62f003b677106286a4a805f"} Dec 03 00:27:15 crc kubenswrapper[4912]: I1203 00:27:15.031627 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pzmmk" podStartSLOduration=6.438517279 podStartE2EDuration="1m3.03160245s" podCreationTimestamp="2025-12-03 00:26:12 +0000 UTC" firstStartedPulling="2025-12-03 00:26:15.399850739 +0000 UTC m=+161.041871289" lastFinishedPulling="2025-12-03 00:27:11.9929359 +0000 UTC m=+217.634956460" observedRunningTime="2025-12-03 00:27:15.029476999 +0000 UTC m=+220.671497579" watchObservedRunningTime="2025-12-03 00:27:15.03160245 +0000 UTC m=+220.673623010" Dec 03 00:27:18 crc kubenswrapper[4912]: I1203 00:27:18.105366 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:27:18 crc kubenswrapper[4912]: I1203 00:27:18.105843 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:27:18 crc kubenswrapper[4912]: I1203 00:27:18.105885 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:27:18 crc kubenswrapper[4912]: I1203 00:27:18.106387 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:27:18 crc kubenswrapper[4912]: I1203 00:27:18.106454 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed" gracePeriod=600 Dec 03 00:27:18 crc kubenswrapper[4912]: I1203 00:27:18.358941 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6c6m2" event={"ID":"da8155c3-0290-44f3-b1da-e95d2dcee000","Type":"ContainerStarted","Data":"6ce1ac9cbf83038a89ef1d372728d58bf30356a3a0cb4e615b5082b7f828d5bd"} Dec 03 00:27:18 crc kubenswrapper[4912]: I1203 00:27:18.361048 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5xn5" event={"ID":"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01","Type":"ContainerStarted","Data":"b439f3237f472dac6b8c966c73d913ddc9319800bc2f97c58c1cf7356c25eec3"} Dec 03 00:27:18 crc kubenswrapper[4912]: I1203 00:27:18.363547 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lqdfv" event={"ID":"9e57c9b1-dd22-4cc5-ac09-f010c8030170","Type":"ContainerStarted","Data":"bfa33806ce33d2f56fb7688c916be7c10a27b158c1a04b671788b58a2615eca8"} Dec 03 00:27:18 crc kubenswrapper[4912]: I1203 00:27:18.379569 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kstcn" event={"ID":"668355d3-1254-4898-a4e9-425b8f480dfd","Type":"ContainerStarted","Data":"c476afc2810b40d75f6454ff0517094ea36b78cb8c8e87332db8def3f936c30d"} Dec 03 00:27:19 crc kubenswrapper[4912]: I1203 00:27:19.639814 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p49z" event={"ID":"f9edbdcf-7866-4d27-a29f-5426596b00db","Type":"ContainerStarted","Data":"0c3620c3e0096d07944633f6a20a319288ca590a5b6225a0f87a20d58b6f1e62"} Dec 03 00:27:19 crc kubenswrapper[4912]: I1203 00:27:19.644709 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed" exitCode=0 Dec 03 00:27:19 crc kubenswrapper[4912]: I1203 00:27:19.644748 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed"} Dec 03 00:27:21 crc kubenswrapper[4912]: I1203 00:27:21.287374 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:27:21 crc kubenswrapper[4912]: I1203 00:27:21.287734 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:27:21 crc kubenswrapper[4912]: I1203 00:27:21.289530 4912 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5vr9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Dec 03 00:27:21 crc kubenswrapper[4912]: I1203 00:27:21.289573 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b5vr9" podUID="61f377a2-06fc-427d-9c3e-38f9b3473e23" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Dec 03 00:27:22 crc kubenswrapper[4912]: I1203 00:27:22.033890 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvg59" event={"ID":"616eb06f-63d7-46e2-a8b0-5162a2e159a2","Type":"ContainerStarted","Data":"c964550d2d04de2254d4d9bc261a3a2158b02f72ea6085276f85abbed6123214"} Dec 03 00:27:23 crc kubenswrapper[4912]: I1203 00:27:23.050593 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:27:23 crc kubenswrapper[4912]: I1203 00:27:23.051372 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:27:23 crc kubenswrapper[4912]: I1203 00:27:23.074913 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"be8ecf588e7c78f1c19a77cfa713fb75d6b341f0fd910a73d8ce9d0ff688932f"} Dec 03 00:27:23 crc kubenswrapper[4912]: I1203 00:27:23.077804 4912 generic.go:334] "Generic (PLEG): container finished" podID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerID="b439f3237f472dac6b8c966c73d913ddc9319800bc2f97c58c1cf7356c25eec3" exitCode=0 Dec 03 00:27:23 crc kubenswrapper[4912]: I1203 00:27:23.078503 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5xn5" event={"ID":"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01","Type":"ContainerDied","Data":"b439f3237f472dac6b8c966c73d913ddc9319800bc2f97c58c1cf7356c25eec3"} Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.108957 4912 generic.go:334] "Generic (PLEG): container finished" podID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerID="bfa33806ce33d2f56fb7688c916be7c10a27b158c1a04b671788b58a2615eca8" exitCode=0 Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.109052 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lqdfv" event={"ID":"9e57c9b1-dd22-4cc5-ac09-f010c8030170","Type":"ContainerDied","Data":"bfa33806ce33d2f56fb7688c916be7c10a27b158c1a04b671788b58a2615eca8"} Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.111613 4912 generic.go:334] "Generic (PLEG): container finished" podID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerID="0c3620c3e0096d07944633f6a20a319288ca590a5b6225a0f87a20d58b6f1e62" exitCode=0 Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.111707 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p49z" event={"ID":"f9edbdcf-7866-4d27-a29f-5426596b00db","Type":"ContainerDied","Data":"0c3620c3e0096d07944633f6a20a319288ca590a5b6225a0f87a20d58b6f1e62"} Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.113547 4912 generic.go:334] "Generic (PLEG): container finished" podID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerID="c964550d2d04de2254d4d9bc261a3a2158b02f72ea6085276f85abbed6123214" exitCode=0 Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.113618 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvg59" event={"ID":"616eb06f-63d7-46e2-a8b0-5162a2e159a2","Type":"ContainerDied","Data":"c964550d2d04de2254d4d9bc261a3a2158b02f72ea6085276f85abbed6123214"} Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.115547 4912 generic.go:334] "Generic (PLEG): container finished" podID="668355d3-1254-4898-a4e9-425b8f480dfd" containerID="c476afc2810b40d75f6454ff0517094ea36b78cb8c8e87332db8def3f936c30d" exitCode=0 Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.115628 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kstcn" event={"ID":"668355d3-1254-4898-a4e9-425b8f480dfd","Type":"ContainerDied","Data":"c476afc2810b40d75f6454ff0517094ea36b78cb8c8e87332db8def3f936c30d"} Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.117262 4912 generic.go:334] "Generic (PLEG): container finished" podID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerID="6ce1ac9cbf83038a89ef1d372728d58bf30356a3a0cb4e615b5082b7f828d5bd" exitCode=0 Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.117292 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6c6m2" event={"ID":"da8155c3-0290-44f3-b1da-e95d2dcee000","Type":"ContainerDied","Data":"6ce1ac9cbf83038a89ef1d372728d58bf30356a3a0cb4e615b5082b7f828d5bd"} Dec 03 00:27:25 crc kubenswrapper[4912]: I1203 00:27:25.138243 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pzmmk" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" containerName="registry-server" probeResult="failure" output=< Dec 03 00:27:25 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 00:27:25 crc kubenswrapper[4912]: > Dec 03 00:27:29 crc kubenswrapper[4912]: I1203 00:27:29.141753 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5xn5" event={"ID":"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01","Type":"ContainerStarted","Data":"e61aae9d4b0fd8677c6c6da3d4fe41bdc81f1a8e6d4608d0e371e975cd68d6d6"} Dec 03 00:27:29 crc kubenswrapper[4912]: I1203 00:27:29.144315 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lqdfv" event={"ID":"9e57c9b1-dd22-4cc5-ac09-f010c8030170","Type":"ContainerStarted","Data":"3b10c29ea464f9e8f4875488d5563f3a94f6a6d4957ba00da6587daddd0903b7"} Dec 03 00:27:29 crc kubenswrapper[4912]: I1203 00:27:29.145731 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf9m5" event={"ID":"179c471f-de9c-4cd6-85ac-b8b3380401ea","Type":"ContainerStarted","Data":"cddddac369d742a7b9a95fa07d4156cfa28dd7e923439d07f752f96b57ef0e8b"} Dec 03 00:27:29 crc kubenswrapper[4912]: I1203 00:27:29.166755 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k5xn5" podStartSLOduration=4.691479898 podStartE2EDuration="1m18.166740256s" podCreationTimestamp="2025-12-03 00:26:11 +0000 UTC" firstStartedPulling="2025-12-03 00:26:14.335709491 +0000 UTC m=+159.977730041" lastFinishedPulling="2025-12-03 00:27:27.810969839 +0000 UTC m=+233.452990399" observedRunningTime="2025-12-03 00:27:29.165094719 +0000 UTC m=+234.807115279" watchObservedRunningTime="2025-12-03 00:27:29.166740256 +0000 UTC m=+234.808760816" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.230769 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-b5vr9" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.417986 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kd5n5"] Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.660378 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p49z" event={"ID":"f9edbdcf-7866-4d27-a29f-5426596b00db","Type":"ContainerStarted","Data":"3cadf8ab231a6629046e59447684578a8c73d7ba4f1a8ba71b93f45cdb736468"} Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.664844 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kstcn" event={"ID":"668355d3-1254-4898-a4e9-425b8f480dfd","Type":"ContainerStarted","Data":"afd134dceef07e8b04b19944ab1af4c151c50289d445bf245297c42fe9126872"} Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.667813 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvg59" event={"ID":"616eb06f-63d7-46e2-a8b0-5162a2e159a2","Type":"ContainerStarted","Data":"155f4fc5456965bfc224b6fe79773d10bc26e0d5edc3260476672fbb6f87a620"} Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.670061 4912 generic.go:334] "Generic (PLEG): container finished" podID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerID="cddddac369d742a7b9a95fa07d4156cfa28dd7e923439d07f752f96b57ef0e8b" exitCode=0 Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.670105 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf9m5" event={"ID":"179c471f-de9c-4cd6-85ac-b8b3380401ea","Type":"ContainerDied","Data":"cddddac369d742a7b9a95fa07d4156cfa28dd7e923439d07f752f96b57ef0e8b"} Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.673695 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6c6m2" event={"ID":"da8155c3-0290-44f3-b1da-e95d2dcee000","Type":"ContainerStarted","Data":"9b0c40b3fdd92cd29e475b2cd59c20b9e8b66f5ec527c61bea323572bc5727b7"} Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.690931 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9p49z" podStartSLOduration=7.127175043 podStartE2EDuration="1m22.690913627s" podCreationTimestamp="2025-12-03 00:26:09 +0000 UTC" firstStartedPulling="2025-12-03 00:26:13.261070779 +0000 UTC m=+158.903091339" lastFinishedPulling="2025-12-03 00:27:28.824809363 +0000 UTC m=+234.466829923" observedRunningTime="2025-12-03 00:27:31.689471695 +0000 UTC m=+237.331492275" watchObservedRunningTime="2025-12-03 00:27:31.690913627 +0000 UTC m=+237.332934187" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.714334 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lqdfv" podStartSLOduration=6.27505503 podStartE2EDuration="1m22.71431467s" podCreationTimestamp="2025-12-03 00:26:09 +0000 UTC" firstStartedPulling="2025-12-03 00:26:12.129694538 +0000 UTC m=+157.771715098" lastFinishedPulling="2025-12-03 00:27:28.568954178 +0000 UTC m=+234.210974738" observedRunningTime="2025-12-03 00:27:31.713583829 +0000 UTC m=+237.355604419" watchObservedRunningTime="2025-12-03 00:27:31.71431467 +0000 UTC m=+237.356335230" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.739697 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6c6m2" podStartSLOduration=6.066175534 podStartE2EDuration="1m22.739676381s" podCreationTimestamp="2025-12-03 00:26:09 +0000 UTC" firstStartedPulling="2025-12-03 00:26:12.093578844 +0000 UTC m=+157.735599404" lastFinishedPulling="2025-12-03 00:27:28.767079691 +0000 UTC m=+234.409100251" observedRunningTime="2025-12-03 00:27:31.738172926 +0000 UTC m=+237.380193496" watchObservedRunningTime="2025-12-03 00:27:31.739676381 +0000 UTC m=+237.381696941" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.777119 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kstcn" podStartSLOduration=6.283203429 podStartE2EDuration="1m19.777080017s" podCreationTimestamp="2025-12-03 00:26:12 +0000 UTC" firstStartedPulling="2025-12-03 00:26:15.396380218 +0000 UTC m=+161.038400768" lastFinishedPulling="2025-12-03 00:27:28.890256796 +0000 UTC m=+234.532277356" observedRunningTime="2025-12-03 00:27:31.776887371 +0000 UTC m=+237.418907951" watchObservedRunningTime="2025-12-03 00:27:31.777080017 +0000 UTC m=+237.419100587" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.804647 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zvg59" podStartSLOduration=6.357337218 podStartE2EDuration="1m20.80462901s" podCreationTimestamp="2025-12-03 00:26:11 +0000 UTC" firstStartedPulling="2025-12-03 00:26:14.309588455 +0000 UTC m=+159.951609015" lastFinishedPulling="2025-12-03 00:27:28.756880247 +0000 UTC m=+234.398900807" observedRunningTime="2025-12-03 00:27:31.801038586 +0000 UTC m=+237.443059146" watchObservedRunningTime="2025-12-03 00:27:31.80462901 +0000 UTC m=+237.446649570" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.894365 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.894447 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.918656 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:27:31 crc kubenswrapper[4912]: I1203 00:27:31.918706 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:27:32 crc kubenswrapper[4912]: I1203 00:27:32.154833 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:27:33 crc kubenswrapper[4912]: I1203 00:27:33.032928 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-zvg59" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerName="registry-server" probeResult="failure" output=< Dec 03 00:27:33 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 00:27:33 crc kubenswrapper[4912]: > Dec 03 00:27:33 crc kubenswrapper[4912]: I1203 00:27:33.062962 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:27:33 crc kubenswrapper[4912]: I1203 00:27:33.126104 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:27:33 crc kubenswrapper[4912]: I1203 00:27:33.192385 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:27:33 crc kubenswrapper[4912]: I1203 00:27:33.192512 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:27:34 crc kubenswrapper[4912]: I1203 00:27:34.235770 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kstcn" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" containerName="registry-server" probeResult="failure" output=< Dec 03 00:27:34 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 00:27:34 crc kubenswrapper[4912]: > Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.538533 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.538839 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.593538 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.607090 4912 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 00:27:39 crc kubenswrapper[4912]: E1203 00:27:39.607362 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ecbc09-cfac-405a-8385-8e2343776d29" containerName="pruner" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.607375 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ecbc09-cfac-405a-8385-8e2343776d29" containerName="pruner" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.607530 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ecbc09-cfac-405a-8385-8e2343776d29" containerName="pruner" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.607969 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.634554 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.688585 4912 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.688894 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0" gracePeriod=15 Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.688938 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9" gracePeriod=15 Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.689023 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91" gracePeriod=15 Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.689060 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8" gracePeriod=15 Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.689108 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12" gracePeriod=15 Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.690811 4912 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 00:27:39 crc kubenswrapper[4912]: E1203 00:27:39.691070 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691080 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:27:39 crc kubenswrapper[4912]: E1203 00:27:39.691090 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691096 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 00:27:39 crc kubenswrapper[4912]: E1203 00:27:39.691110 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691116 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:27:39 crc kubenswrapper[4912]: E1203 00:27:39.691127 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691135 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 00:27:39 crc kubenswrapper[4912]: E1203 00:27:39.691145 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691151 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 00:27:39 crc kubenswrapper[4912]: E1203 00:27:39.691158 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691163 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 00:27:39 crc kubenswrapper[4912]: E1203 00:27:39.691178 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691199 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691296 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691307 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691318 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691326 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691335 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.691345 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.728163 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.728387 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.728833 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.728951 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.729033 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.767236 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.767585 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.813198 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.813843 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.814029 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.814583 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.832693 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.832768 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.832797 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.832856 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.832880 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.832908 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.832936 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.832960 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.833032 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.834210 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.834236 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.835487 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.835577 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.856301 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.856805 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.857085 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.857570 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.857867 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.930664 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.934025 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.934112 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.934141 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.934159 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.934226 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: I1203 00:27:39.934238 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:39 crc kubenswrapper[4912]: E1203 00:27:39.956136 4912 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.20:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d8cfff50f86fb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 00:27:39.955087099 +0000 UTC m=+245.597107659,LastTimestamp:2025-12-03 00:27:39.955087099 +0000 UTC m=+245.597107659,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.191333 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.191710 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.233563 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.234184 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.234713 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.235218 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.235551 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.235878 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.805446 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"6fa3ddf41ac85b18b2abd5695f24c95bbd12957064ec935fd63cdfd178b58077"} Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.844364 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.845071 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.845379 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.845870 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.846084 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.847712 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.848281 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.848611 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.849010 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:40 crc kubenswrapper[4912]: I1203 00:27:40.849269 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: E1203 00:27:41.010645 4912 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: E1203 00:27:41.011063 4912 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: E1203 00:27:41.011449 4912 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: E1203 00:27:41.012037 4912 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: E1203 00:27:41.012465 4912 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.012517 4912 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 00:27:41 crc kubenswrapper[4912]: E1203 00:27:41.012841 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="200ms" Dec 03 00:27:41 crc kubenswrapper[4912]: E1203 00:27:41.213502 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="400ms" Dec 03 00:27:41 crc kubenswrapper[4912]: E1203 00:27:41.614624 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="800ms" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.822244 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.826483 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.827172 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9" exitCode=0 Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.827195 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91" exitCode=0 Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.827205 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12" exitCode=0 Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.827213 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8" exitCode=2 Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.827290 4912 scope.go:117] "RemoveContainer" containerID="e8daa2307681f90d22a996b6d744171a89ec01829c31aae0532d654da84b53f9" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.829399 4912 generic.go:334] "Generic (PLEG): container finished" podID="07a517b0-e70e-4eea-ba9f-3f2b18153008" containerID="7f5a733fb2a6a806700ceee260c0f8b7df28c2735a4609118cff55e5e0e99d32" exitCode=0 Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.832086 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"07a517b0-e70e-4eea-ba9f-3f2b18153008","Type":"ContainerDied","Data":"7f5a733fb2a6a806700ceee260c0f8b7df28c2735a4609118cff55e5e0e99d32"} Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.832780 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.833019 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.833212 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.833410 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.833620 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.848690 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"dca20118b902d37fb9ad3ea25dacc2691dcec43d1c7732a0283b0cbff9e17025"} Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.855550 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.855915 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.856295 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.856588 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.856752 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.942887 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.943595 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.943914 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.944313 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.944509 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.944691 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.944841 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.969294 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.970001 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.970464 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.970830 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.971256 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.971596 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.971916 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:41 crc kubenswrapper[4912]: I1203 00:27:41.972174 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.007395 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.007943 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.008665 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.009187 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.009574 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.009900 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.010244 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.010581 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: E1203 00:27:42.417941 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="1.6s" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.557383 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.559649 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.560309 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.560534 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.560778 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.560960 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.561166 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.561349 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.561711 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.561915 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.671097 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.671165 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.671195 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.671293 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.671370 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.671409 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.671750 4912 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.671773 4912 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.671782 4912 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.858096 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.858749 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0" exitCode=0 Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.858822 4912 scope.go:117] "RemoveContainer" containerID="4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.858926 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.859981 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.860328 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.860584 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.860816 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.861020 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.861248 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.861766 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.862015 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.862245 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf9m5" event={"ID":"179c471f-de9c-4cd6-85ac-b8b3380401ea","Type":"ContainerStarted","Data":"66c92c24baa2245e017a824cc9b5a7d72429c3df4422546d4d78b0ec0065371f"} Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.863555 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.863891 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.864223 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.865321 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.865628 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.866016 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.866486 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.867082 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.867410 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.874182 4912 scope.go:117] "RemoveContainer" containerID="5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.876673 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.879662 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.880155 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.880581 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.880876 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.881333 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.881765 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.882070 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.882975 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.890812 4912 scope.go:117] "RemoveContainer" containerID="daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.940057 4912 scope.go:117] "RemoveContainer" containerID="652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.962765 4912 scope.go:117] "RemoveContainer" containerID="ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0" Dec 03 00:27:42 crc kubenswrapper[4912]: I1203 00:27:42.979086 4912 scope.go:117] "RemoveContainer" containerID="c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.016604 4912 scope.go:117] "RemoveContainer" containerID="4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9" Dec 03 00:27:43 crc kubenswrapper[4912]: E1203 00:27:43.019149 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\": container with ID starting with 4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9 not found: ID does not exist" containerID="4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.019185 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9"} err="failed to get container status \"4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\": rpc error: code = NotFound desc = could not find container \"4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9\": container with ID starting with 4a9e71d0a38a1589ca3d5fda57ca4b740496806c6b9345da831f309672f8edb9 not found: ID does not exist" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.019217 4912 scope.go:117] "RemoveContainer" containerID="5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91" Dec 03 00:27:43 crc kubenswrapper[4912]: E1203 00:27:43.019745 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\": container with ID starting with 5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91 not found: ID does not exist" containerID="5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.019860 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91"} err="failed to get container status \"5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\": rpc error: code = NotFound desc = could not find container \"5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91\": container with ID starting with 5a8ef442f0dc315755c94d8fde40ea371ff266ac06697361b146027ae25dfe91 not found: ID does not exist" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.019884 4912 scope.go:117] "RemoveContainer" containerID="daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12" Dec 03 00:27:43 crc kubenswrapper[4912]: E1203 00:27:43.020222 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\": container with ID starting with daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12 not found: ID does not exist" containerID="daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.020246 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12"} err="failed to get container status \"daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\": rpc error: code = NotFound desc = could not find container \"daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12\": container with ID starting with daf57ecee5f734093499640fc56315dcb2e3573fc77233416b853396d95c1e12 not found: ID does not exist" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.020263 4912 scope.go:117] "RemoveContainer" containerID="652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8" Dec 03 00:27:43 crc kubenswrapper[4912]: E1203 00:27:43.020486 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\": container with ID starting with 652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8 not found: ID does not exist" containerID="652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.020508 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8"} err="failed to get container status \"652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\": rpc error: code = NotFound desc = could not find container \"652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8\": container with ID starting with 652e9187f90360331c4043d2e3f0e19913a6f6932a587f91dab7de5def31fbe8 not found: ID does not exist" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.020525 4912 scope.go:117] "RemoveContainer" containerID="ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0" Dec 03 00:27:43 crc kubenswrapper[4912]: E1203 00:27:43.020765 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\": container with ID starting with ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0 not found: ID does not exist" containerID="ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.020787 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0"} err="failed to get container status \"ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\": rpc error: code = NotFound desc = could not find container \"ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0\": container with ID starting with ba93acdcbfea03f02db3913b1623f8580b23e4ac50c51c42fe8e275a09442eb0 not found: ID does not exist" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.020805 4912 scope.go:117] "RemoveContainer" containerID="c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88" Dec 03 00:27:43 crc kubenswrapper[4912]: E1203 00:27:43.021349 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\": container with ID starting with c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88 not found: ID does not exist" containerID="c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.021376 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88"} err="failed to get container status \"c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\": rpc error: code = NotFound desc = could not find container \"c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88\": container with ID starting with c3875b45aa40df2c2f2a60fe9c513f02803dd1cc157ed23e97459b24ad02ab88 not found: ID does not exist" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.193847 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.194451 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.194737 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.195123 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.195382 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.195679 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.195946 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.196234 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.196421 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.196593 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.242511 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.243185 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.243580 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.243826 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.244085 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.245265 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.245489 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.245709 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.245944 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.246135 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.246298 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.293136 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.294011 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.294362 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.294777 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.295150 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.295529 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.296871 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.297292 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.297619 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.297835 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.298110 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.379524 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-kubelet-dir\") pod \"07a517b0-e70e-4eea-ba9f-3f2b18153008\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.379650 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-var-lock\") pod \"07a517b0-e70e-4eea-ba9f-3f2b18153008\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.379674 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "07a517b0-e70e-4eea-ba9f-3f2b18153008" (UID: "07a517b0-e70e-4eea-ba9f-3f2b18153008"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.379697 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07a517b0-e70e-4eea-ba9f-3f2b18153008-kube-api-access\") pod \"07a517b0-e70e-4eea-ba9f-3f2b18153008\" (UID: \"07a517b0-e70e-4eea-ba9f-3f2b18153008\") " Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.379760 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-var-lock" (OuterVolumeSpecName: "var-lock") pod "07a517b0-e70e-4eea-ba9f-3f2b18153008" (UID: "07a517b0-e70e-4eea-ba9f-3f2b18153008"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.380189 4912 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.380206 4912 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/07a517b0-e70e-4eea-ba9f-3f2b18153008-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.387873 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07a517b0-e70e-4eea-ba9f-3f2b18153008-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "07a517b0-e70e-4eea-ba9f-3f2b18153008" (UID: "07a517b0-e70e-4eea-ba9f-3f2b18153008"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.481409 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/07a517b0-e70e-4eea-ba9f-3f2b18153008-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.869148 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"07a517b0-e70e-4eea-ba9f-3f2b18153008","Type":"ContainerDied","Data":"77517e41cef4bca79468f35377b70f28e519ab681110025418f841467a32c844"} Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.869188 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77517e41cef4bca79468f35377b70f28e519ab681110025418f841467a32c844" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.869245 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.881482 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.881844 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.882033 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.882242 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.882450 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.882632 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.882930 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.883231 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.883491 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: I1203 00:27:43.883734 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:43 crc kubenswrapper[4912]: E1203 00:27:43.987672 4912 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.20:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d8cfff50f86fb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 00:27:39.955087099 +0000 UTC m=+245.597107659,LastTimestamp:2025-12-03 00:27:39.955087099 +0000 UTC m=+245.597107659,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 00:27:44 crc kubenswrapper[4912]: E1203 00:27:44.018882 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="3.2s" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.573578 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.573878 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.574093 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.574465 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.574999 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.575294 4912 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.575595 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.575838 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.576094 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.576351 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:44 crc kubenswrapper[4912]: I1203 00:27:44.578440 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 00:27:47 crc kubenswrapper[4912]: E1203 00:27:47.219863 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="6.4s" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.006915 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.008126 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.046861 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.047405 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.047638 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.047940 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.048398 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.048639 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.048826 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.049003 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.049192 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.049366 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.961864 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.962519 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.962746 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.962920 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.963074 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.963227 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.963472 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.963657 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.963835 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:50 crc kubenswrapper[4912]: I1203 00:27:50.964000 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:51 crc kubenswrapper[4912]: E1203 00:27:51.603120 4912 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.20:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" volumeName="registry-storage" Dec 03 00:27:53 crc kubenswrapper[4912]: E1203 00:27:53.621020 4912 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="7s" Dec 03 00:27:53 crc kubenswrapper[4912]: E1203 00:27:53.989551 4912 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.20:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d8cfff50f86fb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 00:27:39.955087099 +0000 UTC m=+245.597107659,LastTimestamp:2025-12-03 00:27:39.955087099 +0000 UTC m=+245.597107659,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 00:27:54 crc kubenswrapper[4912]: I1203 00:27:54.586321 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:54 crc kubenswrapper[4912]: I1203 00:27:54.588033 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:54 crc kubenswrapper[4912]: I1203 00:27:54.588375 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:54 crc kubenswrapper[4912]: I1203 00:27:54.588715 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:54 crc kubenswrapper[4912]: I1203 00:27:54.589011 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:54 crc kubenswrapper[4912]: I1203 00:27:54.589268 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:54 crc kubenswrapper[4912]: I1203 00:27:54.589578 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:54 crc kubenswrapper[4912]: I1203 00:27:54.589958 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:54 crc kubenswrapper[4912]: I1203 00:27:54.590185 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.571744 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.572958 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.573589 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.574017 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.574415 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.574802 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.575344 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.575878 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.576389 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.577053 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.595598 4912 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.595660 4912 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:27:55 crc kubenswrapper[4912]: E1203 00:27:55.596921 4912 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.598087 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:55 crc kubenswrapper[4912]: W1203 00:27:55.634309 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-e09ba476fa7abc6943c50bec3c10a62a18fff86f30c8dec03bc23a7a5cce892b WatchSource:0}: Error finding container e09ba476fa7abc6943c50bec3c10a62a18fff86f30c8dec03bc23a7a5cce892b: Status 404 returned error can't find the container with id e09ba476fa7abc6943c50bec3c10a62a18fff86f30c8dec03bc23a7a5cce892b Dec 03 00:27:55 crc kubenswrapper[4912]: I1203 00:27:55.958060 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e09ba476fa7abc6943c50bec3c10a62a18fff86f30c8dec03bc23a7a5cce892b"} Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.450964 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" podUID="faa0cccb-7352-4752-8957-0d770dacaee5" containerName="oauth-openshift" containerID="cri-o://161ead0b788d10b8512b924aca6a2951e82533abc1a25202e850c2e717a2f0ea" gracePeriod=15 Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.965026 4912 generic.go:334] "Generic (PLEG): container finished" podID="faa0cccb-7352-4752-8957-0d770dacaee5" containerID="161ead0b788d10b8512b924aca6a2951e82533abc1a25202e850c2e717a2f0ea" exitCode=0 Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.965270 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" event={"ID":"faa0cccb-7352-4752-8957-0d770dacaee5","Type":"ContainerDied","Data":"161ead0b788d10b8512b924aca6a2951e82533abc1a25202e850c2e717a2f0ea"} Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.968786 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.968815 4912 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb" exitCode=1 Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.968850 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb"} Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.969327 4912 scope.go:117] "RemoveContainer" containerID="72e1324565b1b290e8398debc0f72950ba7c98c97218488d85950f1945bc7adb" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.970149 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.970368 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.970658 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.971061 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.971290 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.971707 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.971970 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.972383 4912 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="9be1dcadf72e39a30383eb79695619be2eeba64505499909d406d52aa1ba3f33" exitCode=0 Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.972416 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"9be1dcadf72e39a30383eb79695619be2eeba64505499909d406d52aa1ba3f33"} Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.972644 4912 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.972655 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.972671 4912 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.972926 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: E1203 00:27:56.972932 4912 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.973112 4912 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.973367 4912 status_manager.go:851] "Failed to get status for pod" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" pod="openshift-marketplace/certified-operators-rf9m5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-rf9m5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.973618 4912 status_manager.go:851] "Failed to get status for pod" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" pod="openshift-marketplace/redhat-operators-kstcn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kstcn\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.973859 4912 status_manager.go:851] "Failed to get status for pod" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.974090 4912 status_manager.go:851] "Failed to get status for pod" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" pod="openshift-marketplace/redhat-marketplace-zvg59" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-zvg59\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.974364 4912 status_manager.go:851] "Failed to get status for pod" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" pod="openshift-marketplace/certified-operators-lqdfv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-lqdfv\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.974716 4912 status_manager.go:851] "Failed to get status for pod" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" pod="openshift-marketplace/redhat-marketplace-k5xn5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-k5xn5\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.975230 4912 status_manager.go:851] "Failed to get status for pod" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" pod="openshift-marketplace/community-operators-6c6m2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-6c6m2\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.975943 4912 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.976183 4912 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:56 crc kubenswrapper[4912]: I1203 00:27:56.976386 4912 status_manager.go:851] "Failed to get status for pod" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" pod="openshift-marketplace/community-operators-9p49z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9p49z\": dial tcp 38.102.83.20:6443: connect: connection refused" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.450744 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493509 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-router-certs\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493556 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-ocp-branding-template\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493582 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/faa0cccb-7352-4752-8957-0d770dacaee5-audit-dir\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493611 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-provider-selection\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493634 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-trusted-ca-bundle\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493653 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-error\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493676 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-cliconfig\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493725 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-session\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493754 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-serving-cert\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493773 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-service-ca\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493799 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-login\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493833 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-audit-policies\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493853 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g9v6\" (UniqueName: \"kubernetes.io/projected/faa0cccb-7352-4752-8957-0d770dacaee5-kube-api-access-4g9v6\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.493872 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-idp-0-file-data\") pod \"faa0cccb-7352-4752-8957-0d770dacaee5\" (UID: \"faa0cccb-7352-4752-8957-0d770dacaee5\") " Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.495605 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.499575 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.499646 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.503008 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.503135 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.503439 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.503551 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/faa0cccb-7352-4752-8957-0d770dacaee5-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.497082 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.503610 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.503657 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faa0cccb-7352-4752-8957-0d770dacaee5-kube-api-access-4g9v6" (OuterVolumeSpecName: "kube-api-access-4g9v6") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "kube-api-access-4g9v6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.503796 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.503937 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.515168 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.516100 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "faa0cccb-7352-4752-8957-0d770dacaee5" (UID: "faa0cccb-7352-4752-8957-0d770dacaee5"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.595515 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.595999 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596017 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596034 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596049 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596063 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596080 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596095 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596111 4912 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/faa0cccb-7352-4752-8957-0d770dacaee5-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596124 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g9v6\" (UniqueName: \"kubernetes.io/projected/faa0cccb-7352-4752-8957-0d770dacaee5-kube-api-access-4g9v6\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596138 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596150 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596164 4912 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/faa0cccb-7352-4752-8957-0d770dacaee5-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.596178 4912 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/faa0cccb-7352-4752-8957-0d770dacaee5-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.629088 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.982538 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.982649 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c30b35290699a9dd97a1c31fb8e30f4b33faca4ca39ebce7cbf7c4311fbe1f2a"} Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.989645 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"576812dcff887f4e9a465934624e5b5fb922248a4f2c9e41d038d1fc78410b88"} Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.989685 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"21ac9ce74e00ba4e72157f236e5161833ea8fb247da074499e0fa9d1943eb240"} Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.989699 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4b2b571ab611dea6636cf62497d78e34ba3b5d558603b3c848a85bebc41002ed"} Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.991784 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" event={"ID":"faa0cccb-7352-4752-8957-0d770dacaee5","Type":"ContainerDied","Data":"7c010b59ce3b3661f1e87ec3050603cbc3f7e608d37f407cf1e5ce4ee0b7912d"} Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.991822 4912 scope.go:117] "RemoveContainer" containerID="161ead0b788d10b8512b924aca6a2951e82533abc1a25202e850c2e717a2f0ea" Dec 03 00:27:57 crc kubenswrapper[4912]: I1203 00:27:57.991952 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kd5n5" Dec 03 00:27:59 crc kubenswrapper[4912]: I1203 00:27:59.001800 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c59beaa0506b6e06cd49d834eb37b248c36d14c72a8fea7040076541a2d7c839"} Dec 03 00:27:59 crc kubenswrapper[4912]: I1203 00:27:59.002141 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"761ac1026adf51038a729bb48836ba60ca98abf97f3116e2929671bc8dd1d68d"} Dec 03 00:27:59 crc kubenswrapper[4912]: I1203 00:27:59.002064 4912 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:27:59 crc kubenswrapper[4912]: I1203 00:27:59.002194 4912 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:27:59 crc kubenswrapper[4912]: I1203 00:27:59.002168 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:28:00 crc kubenswrapper[4912]: I1203 00:28:00.230045 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:28:00 crc kubenswrapper[4912]: I1203 00:28:00.230295 4912 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 00:28:00 crc kubenswrapper[4912]: I1203 00:28:00.230350 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 00:28:00 crc kubenswrapper[4912]: I1203 00:28:00.599367 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:28:00 crc kubenswrapper[4912]: I1203 00:28:00.599447 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:28:00 crc kubenswrapper[4912]: I1203 00:28:00.604673 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:28:04 crc kubenswrapper[4912]: I1203 00:28:04.023233 4912 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:28:04 crc kubenswrapper[4912]: I1203 00:28:04.603875 4912 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="68604374-b585-4bb6-be10-22c002f2e477" Dec 03 00:28:05 crc kubenswrapper[4912]: I1203 00:28:05.049070 4912 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:28:05 crc kubenswrapper[4912]: I1203 00:28:05.049106 4912 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:28:05 crc kubenswrapper[4912]: I1203 00:28:05.052869 4912 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="68604374-b585-4bb6-be10-22c002f2e477" Dec 03 00:28:05 crc kubenswrapper[4912]: I1203 00:28:05.054204 4912 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://4b2b571ab611dea6636cf62497d78e34ba3b5d558603b3c848a85bebc41002ed" Dec 03 00:28:05 crc kubenswrapper[4912]: I1203 00:28:05.054244 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:28:06 crc kubenswrapper[4912]: I1203 00:28:06.054453 4912 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:28:06 crc kubenswrapper[4912]: I1203 00:28:06.054481 4912 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:28:06 crc kubenswrapper[4912]: I1203 00:28:06.058505 4912 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="68604374-b585-4bb6-be10-22c002f2e477" Dec 03 00:28:07 crc kubenswrapper[4912]: I1203 00:28:07.628733 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:28:10 crc kubenswrapper[4912]: I1203 00:28:10.233861 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:28:10 crc kubenswrapper[4912]: I1203 00:28:10.238753 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 00:28:14 crc kubenswrapper[4912]: I1203 00:28:14.080727 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 00:28:14 crc kubenswrapper[4912]: I1203 00:28:14.299150 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 00:28:14 crc kubenswrapper[4912]: I1203 00:28:14.349253 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 00:28:14 crc kubenswrapper[4912]: I1203 00:28:14.565853 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 00:28:14 crc kubenswrapper[4912]: I1203 00:28:14.940392 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 00:28:15 crc kubenswrapper[4912]: I1203 00:28:15.191927 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 00:28:15 crc kubenswrapper[4912]: I1203 00:28:15.624972 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 00:28:15 crc kubenswrapper[4912]: I1203 00:28:15.745631 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 00:28:15 crc kubenswrapper[4912]: I1203 00:28:15.962555 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 00:28:15 crc kubenswrapper[4912]: I1203 00:28:15.962888 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 00:28:15 crc kubenswrapper[4912]: I1203 00:28:15.997095 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.009953 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.030724 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.321481 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.393975 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.432358 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.511310 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.695814 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.698542 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.741000 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.761968 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.849048 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 00:28:16 crc kubenswrapper[4912]: I1203 00:28:16.963498 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.017301 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.056816 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.063842 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.157134 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.188363 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.285123 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.290913 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.296158 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.379081 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.421637 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.426520 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.537005 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.564582 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.662921 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.703289 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 00:28:17 crc kubenswrapper[4912]: I1203 00:28:17.969981 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.308801 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.338502 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.345643 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.392080 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.395858 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.475399 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.582911 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.587516 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.729521 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.772878 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.795323 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.867170 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 00:28:18 crc kubenswrapper[4912]: I1203 00:28:18.969814 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.098723 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.105201 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.112289 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.152162 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.153208 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.215972 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.292574 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.348016 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.363504 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.384168 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.399956 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.473059 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.592774 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.599396 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.686406 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.801390 4912 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.802931 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rf9m5" podStartSLOduration=41.689424995 podStartE2EDuration="2m10.802903618s" podCreationTimestamp="2025-12-03 00:26:09 +0000 UTC" firstStartedPulling="2025-12-03 00:26:13.288362899 +0000 UTC m=+158.930383459" lastFinishedPulling="2025-12-03 00:27:42.401841522 +0000 UTC m=+248.043862082" observedRunningTime="2025-12-03 00:28:03.982928366 +0000 UTC m=+269.624948926" watchObservedRunningTime="2025-12-03 00:28:19.802903618 +0000 UTC m=+285.444924178" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.804893 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=40.804885075 podStartE2EDuration="40.804885075s" podCreationTimestamp="2025-12-03 00:27:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:28:03.917082523 +0000 UTC m=+269.559103083" watchObservedRunningTime="2025-12-03 00:28:19.804885075 +0000 UTC m=+285.446905635" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.807662 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kd5n5","openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.807733 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-6f89645989-7t5zq"] Dec 03 00:28:19 crc kubenswrapper[4912]: E1203 00:28:19.808270 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faa0cccb-7352-4752-8957-0d770dacaee5" containerName="oauth-openshift" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.808312 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="faa0cccb-7352-4752-8957-0d770dacaee5" containerName="oauth-openshift" Dec 03 00:28:19 crc kubenswrapper[4912]: E1203 00:28:19.808337 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" containerName="installer" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.808369 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" containerName="installer" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.808368 4912 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.808419 4912 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3bf502f4-28d8-4d3e-bbfc-ebd54758b252" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.808496 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="faa0cccb-7352-4752-8957-0d770dacaee5" containerName="oauth-openshift" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.808510 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="07a517b0-e70e-4eea-ba9f-3f2b18153008" containerName="installer" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.809117 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.813179 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.816799 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.817035 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.817364 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.817378 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.818103 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.818287 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.818405 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.818486 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.818692 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.818778 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.819627 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.821894 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/266fc078-be08-4cad-a9d6-2e6440bdd1a6-audit-dir\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.821946 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-template-login\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.821988 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822020 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mltgb\" (UniqueName: \"kubernetes.io/projected/266fc078-be08-4cad-a9d6-2e6440bdd1a6-kube-api-access-mltgb\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822039 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822057 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822078 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822107 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-template-error\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822130 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-audit-policies\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822149 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822164 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822195 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822214 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-session\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.822238 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.824127 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.828859 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.830641 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.834625 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.868950 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.868920366 podStartE2EDuration="15.868920366s" podCreationTimestamp="2025-12-03 00:28:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:28:19.860752241 +0000 UTC m=+285.502772821" watchObservedRunningTime="2025-12-03 00:28:19.868920366 +0000 UTC m=+285.510940936" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.869853 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.923740 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.923807 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-session\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.923852 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.923884 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/266fc078-be08-4cad-a9d6-2e6440bdd1a6-audit-dir\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.923907 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-template-login\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.923933 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.923966 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mltgb\" (UniqueName: \"kubernetes.io/projected/266fc078-be08-4cad-a9d6-2e6440bdd1a6-kube-api-access-mltgb\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.923982 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.924001 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.924020 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.924052 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-template-error\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.924075 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-audit-policies\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.924098 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.924118 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.924569 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.925047 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-service-ca\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.925976 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-audit-policies\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.926375 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.927056 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/266fc078-be08-4cad-a9d6-2e6440bdd1a6-audit-dir\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.933180 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-router-certs\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.934217 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-template-error\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.934582 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.935450 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.945081 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-template-login\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.949162 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mltgb\" (UniqueName: \"kubernetes.io/projected/266fc078-be08-4cad-a9d6-2e6440bdd1a6-kube-api-access-mltgb\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.949313 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-session\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.955590 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.961443 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/266fc078-be08-4cad-a9d6-2e6440bdd1a6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6f89645989-7t5zq\" (UID: \"266fc078-be08-4cad-a9d6-2e6440bdd1a6\") " pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:19 crc kubenswrapper[4912]: I1203 00:28:19.982792 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.086776 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.129719 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.204389 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.216469 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.264533 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.321580 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.395984 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.424380 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.460885 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.505352 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.528847 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6f89645989-7t5zq"] Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.550696 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.581588 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faa0cccb-7352-4752-8957-0d770dacaee5" path="/var/lib/kubelet/pods/faa0cccb-7352-4752-8957-0d770dacaee5/volumes" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.653248 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.697735 4912 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.720972 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.847720 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.861552 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.866710 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.883406 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 00:28:20 crc kubenswrapper[4912]: I1203 00:28:20.898673 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.000150 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.014180 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.045841 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.258222 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.343619 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.344205 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.344392 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.435071 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.506553 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.512571 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.520304 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.637843 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.723320 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.806800 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.820678 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 00:28:21 crc kubenswrapper[4912]: I1203 00:28:21.841956 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.095121 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.162396 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.200655 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.253806 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.342106 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.393320 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.430092 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.475570 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.499565 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.531776 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.590530 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.661394 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.766695 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.768386 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.788820 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 00:28:22 crc kubenswrapper[4912]: I1203 00:28:22.849780 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.105508 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.252180 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 00:28:23 crc kubenswrapper[4912]: E1203 00:28:23.283066 4912 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 03 00:28:23 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3" Netns:"/var/run/netns/8f665cd4-d141-4613-b1ea-3e803d52e424" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod "oauth-openshift-6f89645989-7t5zq" not found Dec 03 00:28:23 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:28:23 crc kubenswrapper[4912]: > Dec 03 00:28:23 crc kubenswrapper[4912]: E1203 00:28:23.283193 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 03 00:28:23 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3" Netns:"/var/run/netns/8f665cd4-d141-4613-b1ea-3e803d52e424" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod "oauth-openshift-6f89645989-7t5zq" not found Dec 03 00:28:23 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:28:23 crc kubenswrapper[4912]: > pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:23 crc kubenswrapper[4912]: E1203 00:28:23.283235 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 03 00:28:23 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3" Netns:"/var/run/netns/8f665cd4-d141-4613-b1ea-3e803d52e424" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod "oauth-openshift-6f89645989-7t5zq" not found Dec 03 00:28:23 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:28:23 crc kubenswrapper[4912]: > pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:23 crc kubenswrapper[4912]: E1203 00:28:23.283320 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-6f89645989-7t5zq_openshift-authentication(266fc078-be08-4cad-a9d6-2e6440bdd1a6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-6f89645989-7t5zq_openshift-authentication(266fc078-be08-4cad-a9d6-2e6440bdd1a6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3\\\" Netns:\\\"/var/run/netns/8f665cd4-d141-4613-b1ea-3e803d52e424\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=df3754dd505d6d6e8d35b00ab02814ff55332ee99d93417e364c9293d67de8f3;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod \\\"oauth-openshift-6f89645989-7t5zq\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" podUID="266fc078-be08-4cad-a9d6-2e6440bdd1a6" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.334002 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.378964 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.407949 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.499280 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.557052 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.567953 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.625514 4912 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.782232 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.783213 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.850661 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 00:28:23 crc kubenswrapper[4912]: I1203 00:28:23.958227 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.046631 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.111867 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.141143 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.144625 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.166410 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.167303 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.178292 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.318815 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.402991 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.413259 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.482743 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.622857 4912 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.793541 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.800647 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 00:28:24 crc kubenswrapper[4912]: I1203 00:28:24.907053 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.030797 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.096478 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.141563 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.185577 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.218853 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.279898 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.332227 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.369278 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.435724 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.480962 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.493613 4912 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.499623 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.500903 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.514330 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.556565 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.590081 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.596299 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.666287 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.684101 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.723932 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.730160 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.815681 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 00:28:25 crc kubenswrapper[4912]: I1203 00:28:25.906548 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.021936 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.062220 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.108711 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.121723 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.136496 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.238500 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.319709 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.466028 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.586545 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.602009 4912 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.602249 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://dca20118b902d37fb9ad3ea25dacc2691dcec43d1c7732a0283b0cbff9e17025" gracePeriod=5 Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.671794 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.698059 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.771778 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.854655 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 00:28:26 crc kubenswrapper[4912]: I1203 00:28:26.952353 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 00:28:27 crc kubenswrapper[4912]: I1203 00:28:27.021718 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 00:28:27 crc kubenswrapper[4912]: I1203 00:28:27.127228 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 00:28:27 crc kubenswrapper[4912]: I1203 00:28:27.231459 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 00:28:27 crc kubenswrapper[4912]: E1203 00:28:27.320092 4912 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 03 00:28:27 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f" Netns:"/var/run/netns/e8683eaf-ef65-47ae-bfbe-175374026015" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod "oauth-openshift-6f89645989-7t5zq" not found Dec 03 00:28:27 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:28:27 crc kubenswrapper[4912]: > Dec 03 00:28:27 crc kubenswrapper[4912]: E1203 00:28:27.320164 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 03 00:28:27 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f" Netns:"/var/run/netns/e8683eaf-ef65-47ae-bfbe-175374026015" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod "oauth-openshift-6f89645989-7t5zq" not found Dec 03 00:28:27 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:28:27 crc kubenswrapper[4912]: > pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:27 crc kubenswrapper[4912]: E1203 00:28:27.320186 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 03 00:28:27 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f" Netns:"/var/run/netns/e8683eaf-ef65-47ae-bfbe-175374026015" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod "oauth-openshift-6f89645989-7t5zq" not found Dec 03 00:28:27 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:28:27 crc kubenswrapper[4912]: > pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:27 crc kubenswrapper[4912]: E1203 00:28:27.320240 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-6f89645989-7t5zq_openshift-authentication(266fc078-be08-4cad-a9d6-2e6440bdd1a6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-6f89645989-7t5zq_openshift-authentication(266fc078-be08-4cad-a9d6-2e6440bdd1a6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f\\\" Netns:\\\"/var/run/netns/e8683eaf-ef65-47ae-bfbe-175374026015\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=f130e29afdad33d5e5d6c8daf630569bc379925c6c9bccf0e270b54d6135af4f;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod \\\"oauth-openshift-6f89645989-7t5zq\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" podUID="266fc078-be08-4cad-a9d6-2e6440bdd1a6" Dec 03 00:28:27 crc kubenswrapper[4912]: I1203 00:28:27.388804 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 00:28:27 crc kubenswrapper[4912]: I1203 00:28:27.512229 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 00:28:27 crc kubenswrapper[4912]: I1203 00:28:27.828997 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 00:28:27 crc kubenswrapper[4912]: I1203 00:28:27.945461 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 00:28:28 crc kubenswrapper[4912]: I1203 00:28:28.154353 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 00:28:28 crc kubenswrapper[4912]: I1203 00:28:28.172040 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 00:28:28 crc kubenswrapper[4912]: I1203 00:28:28.213829 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 00:28:28 crc kubenswrapper[4912]: I1203 00:28:28.835383 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 00:28:28 crc kubenswrapper[4912]: I1203 00:28:28.924506 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 00:28:28 crc kubenswrapper[4912]: I1203 00:28:28.958465 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 00:28:28 crc kubenswrapper[4912]: I1203 00:28:28.992610 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 00:28:29 crc kubenswrapper[4912]: I1203 00:28:29.027016 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 00:28:29 crc kubenswrapper[4912]: I1203 00:28:29.040448 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 00:28:29 crc kubenswrapper[4912]: I1203 00:28:29.082847 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 00:28:29 crc kubenswrapper[4912]: I1203 00:28:29.115673 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 00:28:29 crc kubenswrapper[4912]: I1203 00:28:29.147186 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 00:28:29 crc kubenswrapper[4912]: I1203 00:28:29.586967 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 00:28:29 crc kubenswrapper[4912]: I1203 00:28:29.601794 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 00:28:29 crc kubenswrapper[4912]: I1203 00:28:29.817159 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 00:28:30 crc kubenswrapper[4912]: I1203 00:28:30.441780 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.345114 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.345382 4912 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="dca20118b902d37fb9ad3ea25dacc2691dcec43d1c7732a0283b0cbff9e17025" exitCode=137 Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.769241 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.769590 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.878872 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.879192 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.879302 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.878981 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.879328 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.879419 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.879458 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.879631 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.879878 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.880245 4912 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.880263 4912 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.880272 4912 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.880281 4912 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.887934 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:28:32 crc kubenswrapper[4912]: I1203 00:28:32.981723 4912 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 00:28:33 crc kubenswrapper[4912]: I1203 00:28:33.352170 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 00:28:33 crc kubenswrapper[4912]: I1203 00:28:33.352244 4912 scope.go:117] "RemoveContainer" containerID="dca20118b902d37fb9ad3ea25dacc2691dcec43d1c7732a0283b0cbff9e17025" Dec 03 00:28:33 crc kubenswrapper[4912]: I1203 00:28:33.352315 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 00:28:34 crc kubenswrapper[4912]: I1203 00:28:34.597236 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 00:28:34 crc kubenswrapper[4912]: I1203 00:28:34.597788 4912 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 03 00:28:34 crc kubenswrapper[4912]: I1203 00:28:34.612788 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 00:28:34 crc kubenswrapper[4912]: I1203 00:28:34.612858 4912 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="9d54135c-b939-46f9-a0f8-847bf2d4e535" Dec 03 00:28:34 crc kubenswrapper[4912]: I1203 00:28:34.615938 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 00:28:34 crc kubenswrapper[4912]: I1203 00:28:34.615976 4912 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="9d54135c-b939-46f9-a0f8-847bf2d4e535" Dec 03 00:28:38 crc kubenswrapper[4912]: I1203 00:28:38.571008 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:38 crc kubenswrapper[4912]: I1203 00:28:38.571618 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:38 crc kubenswrapper[4912]: I1203 00:28:38.955700 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 00:28:40 crc kubenswrapper[4912]: I1203 00:28:40.078899 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 00:28:41 crc kubenswrapper[4912]: I1203 00:28:41.330875 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 00:28:41 crc kubenswrapper[4912]: I1203 00:28:41.399255 4912 generic.go:334] "Generic (PLEG): container finished" podID="c916e93d-66d8-436b-841e-c7a92f823912" containerID="acb670445e0d86c859c31d49fe0734cd7907ebf6d9586c532ee7efaea4e0c711" exitCode=0 Dec 03 00:28:41 crc kubenswrapper[4912]: I1203 00:28:41.399296 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" event={"ID":"c916e93d-66d8-436b-841e-c7a92f823912","Type":"ContainerDied","Data":"acb670445e0d86c859c31d49fe0734cd7907ebf6d9586c532ee7efaea4e0c711"} Dec 03 00:28:41 crc kubenswrapper[4912]: I1203 00:28:41.399764 4912 scope.go:117] "RemoveContainer" containerID="acb670445e0d86c859c31d49fe0734cd7907ebf6d9586c532ee7efaea4e0c711" Dec 03 00:28:41 crc kubenswrapper[4912]: E1203 00:28:41.399995 4912 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 03 00:28:41 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be" Netns:"/var/run/netns/6c7e0539-05d4-427f-92c8-fd7bd0c18479" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod "oauth-openshift-6f89645989-7t5zq" not found Dec 03 00:28:41 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:28:41 crc kubenswrapper[4912]: > Dec 03 00:28:41 crc kubenswrapper[4912]: E1203 00:28:41.400052 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 03 00:28:41 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be" Netns:"/var/run/netns/6c7e0539-05d4-427f-92c8-fd7bd0c18479" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod "oauth-openshift-6f89645989-7t5zq" not found Dec 03 00:28:41 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:28:41 crc kubenswrapper[4912]: > pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:41 crc kubenswrapper[4912]: E1203 00:28:41.400075 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 03 00:28:41 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be" Netns:"/var/run/netns/6c7e0539-05d4-427f-92c8-fd7bd0c18479" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod "oauth-openshift-6f89645989-7t5zq" not found Dec 03 00:28:41 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:28:41 crc kubenswrapper[4912]: > pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:41 crc kubenswrapper[4912]: E1203 00:28:41.400162 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-6f89645989-7t5zq_openshift-authentication(266fc078-be08-4cad-a9d6-2e6440bdd1a6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-6f89645989-7t5zq_openshift-authentication(266fc078-be08-4cad-a9d6-2e6440bdd1a6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-6f89645989-7t5zq_openshift-authentication_266fc078-be08-4cad-a9d6-2e6440bdd1a6_0(5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be): error adding pod openshift-authentication_oauth-openshift-6f89645989-7t5zq to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be\\\" Netns:\\\"/var/run/netns/6c7e0539-05d4-427f-92c8-fd7bd0c18479\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-6f89645989-7t5zq;K8S_POD_INFRA_CONTAINER_ID=5a2927b75ef42e847a2e7758900450f4eebdf357b775cdfb4bcacf63b7e489be;K8S_POD_UID=266fc078-be08-4cad-a9d6-2e6440bdd1a6\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-6f89645989-7t5zq] networking: Multus: [openshift-authentication/oauth-openshift-6f89645989-7t5zq/266fc078-be08-4cad-a9d6-2e6440bdd1a6]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-6f89645989-7t5zq in out of cluster comm: pod \\\"oauth-openshift-6f89645989-7t5zq\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" podUID="266fc078-be08-4cad-a9d6-2e6440bdd1a6" Dec 03 00:28:42 crc kubenswrapper[4912]: I1203 00:28:42.407455 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" event={"ID":"c916e93d-66d8-436b-841e-c7a92f823912","Type":"ContainerStarted","Data":"c7c24a61b53aa88d2ec8b06474ee499c238c8210ae197b0ce556db6635edaaaf"} Dec 03 00:28:42 crc kubenswrapper[4912]: I1203 00:28:42.408094 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:28:42 crc kubenswrapper[4912]: I1203 00:28:42.409624 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:28:43 crc kubenswrapper[4912]: I1203 00:28:43.201576 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 00:28:45 crc kubenswrapper[4912]: I1203 00:28:45.459723 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 00:28:47 crc kubenswrapper[4912]: I1203 00:28:47.715539 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 00:28:49 crc kubenswrapper[4912]: I1203 00:28:49.703906 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 00:28:50 crc kubenswrapper[4912]: I1203 00:28:50.177483 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 00:28:50 crc kubenswrapper[4912]: I1203 00:28:50.363275 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 00:28:50 crc kubenswrapper[4912]: I1203 00:28:50.561462 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 00:28:52 crc kubenswrapper[4912]: I1203 00:28:52.829336 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 00:28:53 crc kubenswrapper[4912]: I1203 00:28:53.868490 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 00:28:55 crc kubenswrapper[4912]: I1203 00:28:55.066124 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 00:28:55 crc kubenswrapper[4912]: I1203 00:28:55.157854 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 00:28:55 crc kubenswrapper[4912]: I1203 00:28:55.571396 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:55 crc kubenswrapper[4912]: I1203 00:28:55.572052 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:55 crc kubenswrapper[4912]: I1203 00:28:55.623160 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 00:28:55 crc kubenswrapper[4912]: I1203 00:28:55.759720 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6f89645989-7t5zq"] Dec 03 00:28:56 crc kubenswrapper[4912]: I1203 00:28:56.486262 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" event={"ID":"266fc078-be08-4cad-a9d6-2e6440bdd1a6","Type":"ContainerStarted","Data":"ffb18c758b6a4b31e369e769fdc1934a7151630bae091c3be36fddba58d5673a"} Dec 03 00:28:56 crc kubenswrapper[4912]: I1203 00:28:56.486569 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" event={"ID":"266fc078-be08-4cad-a9d6-2e6440bdd1a6","Type":"ContainerStarted","Data":"83c7f7a29d25d94fa63d052a7d907e3cfc8616f01ad18266868071353412301e"} Dec 03 00:28:56 crc kubenswrapper[4912]: I1203 00:28:56.488985 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:56 crc kubenswrapper[4912]: I1203 00:28:56.510506 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" podStartSLOduration=85.510488565 podStartE2EDuration="1m25.510488565s" podCreationTimestamp="2025-12-03 00:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:28:56.506887681 +0000 UTC m=+322.148908261" watchObservedRunningTime="2025-12-03 00:28:56.510488565 +0000 UTC m=+322.152509125" Dec 03 00:28:56 crc kubenswrapper[4912]: I1203 00:28:56.748465 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 00:28:56 crc kubenswrapper[4912]: I1203 00:28:56.853862 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6f89645989-7t5zq" Dec 03 00:28:57 crc kubenswrapper[4912]: I1203 00:28:57.005618 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 00:28:57 crc kubenswrapper[4912]: I1203 00:28:57.093739 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 00:28:59 crc kubenswrapper[4912]: I1203 00:28:59.384388 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 00:29:00 crc kubenswrapper[4912]: I1203 00:29:00.176647 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 00:29:00 crc kubenswrapper[4912]: I1203 00:29:00.927175 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 00:29:01 crc kubenswrapper[4912]: I1203 00:29:01.825899 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7x9ms"] Dec 03 00:29:01 crc kubenswrapper[4912]: I1203 00:29:01.826145 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" podUID="f33e67e7-af60-4c02-86f0-1bfd3de3d690" containerName="controller-manager" containerID="cri-o://febe37126fabf8ec245b1ec2adf967cc43b2ae15a06c8444cef431b5bac71e6b" gracePeriod=30 Dec 03 00:29:01 crc kubenswrapper[4912]: I1203 00:29:01.940772 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk"] Dec 03 00:29:01 crc kubenswrapper[4912]: I1203 00:29:01.941022 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" podUID="d00bf964-db4a-449d-ad97-d563b4b70f50" containerName="route-controller-manager" containerID="cri-o://a99002b2a1d9a859d66bdc1b157277ad9fe2c114d9fbf1cb5d519cd3eeb2506f" gracePeriod=30 Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.051601 4912 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-7x9ms container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.051962 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" podUID="f33e67e7-af60-4c02-86f0-1bfd3de3d690" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.131812 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.222850 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.277529 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.528308 4912 generic.go:334] "Generic (PLEG): container finished" podID="d00bf964-db4a-449d-ad97-d563b4b70f50" containerID="a99002b2a1d9a859d66bdc1b157277ad9fe2c114d9fbf1cb5d519cd3eeb2506f" exitCode=0 Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.528394 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" event={"ID":"d00bf964-db4a-449d-ad97-d563b4b70f50","Type":"ContainerDied","Data":"a99002b2a1d9a859d66bdc1b157277ad9fe2c114d9fbf1cb5d519cd3eeb2506f"} Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.529898 4912 generic.go:334] "Generic (PLEG): container finished" podID="f33e67e7-af60-4c02-86f0-1bfd3de3d690" containerID="febe37126fabf8ec245b1ec2adf967cc43b2ae15a06c8444cef431b5bac71e6b" exitCode=0 Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.529929 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" event={"ID":"f33e67e7-af60-4c02-86f0-1bfd3de3d690","Type":"ContainerDied","Data":"febe37126fabf8ec245b1ec2adf967cc43b2ae15a06c8444cef431b5bac71e6b"} Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.685689 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.772636 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.781336 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.865375 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f33e67e7-af60-4c02-86f0-1bfd3de3d690-serving-cert\") pod \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.865484 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-proxy-ca-bundles\") pod \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.865513 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-client-ca\") pod \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.865534 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qp8j\" (UniqueName: \"kubernetes.io/projected/f33e67e7-af60-4c02-86f0-1bfd3de3d690-kube-api-access-7qp8j\") pod \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.865564 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-config\") pod \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\" (UID: \"f33e67e7-af60-4c02-86f0-1bfd3de3d690\") " Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.866398 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-client-ca" (OuterVolumeSpecName: "client-ca") pod "f33e67e7-af60-4c02-86f0-1bfd3de3d690" (UID: "f33e67e7-af60-4c02-86f0-1bfd3de3d690"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.866669 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-config" (OuterVolumeSpecName: "config") pod "f33e67e7-af60-4c02-86f0-1bfd3de3d690" (UID: "f33e67e7-af60-4c02-86f0-1bfd3de3d690"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.866780 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f33e67e7-af60-4c02-86f0-1bfd3de3d690" (UID: "f33e67e7-af60-4c02-86f0-1bfd3de3d690"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.870966 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f33e67e7-af60-4c02-86f0-1bfd3de3d690-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f33e67e7-af60-4c02-86f0-1bfd3de3d690" (UID: "f33e67e7-af60-4c02-86f0-1bfd3de3d690"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.871544 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f33e67e7-af60-4c02-86f0-1bfd3de3d690-kube-api-access-7qp8j" (OuterVolumeSpecName: "kube-api-access-7qp8j") pod "f33e67e7-af60-4c02-86f0-1bfd3de3d690" (UID: "f33e67e7-af60-4c02-86f0-1bfd3de3d690"). InnerVolumeSpecName "kube-api-access-7qp8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.967019 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g9n9\" (UniqueName: \"kubernetes.io/projected/d00bf964-db4a-449d-ad97-d563b4b70f50-kube-api-access-6g9n9\") pod \"d00bf964-db4a-449d-ad97-d563b4b70f50\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.967452 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-client-ca\") pod \"d00bf964-db4a-449d-ad97-d563b4b70f50\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.967599 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-config\") pod \"d00bf964-db4a-449d-ad97-d563b4b70f50\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.967681 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d00bf964-db4a-449d-ad97-d563b4b70f50-serving-cert\") pod \"d00bf964-db4a-449d-ad97-d563b4b70f50\" (UID: \"d00bf964-db4a-449d-ad97-d563b4b70f50\") " Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.967972 4912 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.967987 4912 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.967999 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qp8j\" (UniqueName: \"kubernetes.io/projected/f33e67e7-af60-4c02-86f0-1bfd3de3d690-kube-api-access-7qp8j\") on node \"crc\" DevicePath \"\"" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.968015 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e67e7-af60-4c02-86f0-1bfd3de3d690-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.968027 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f33e67e7-af60-4c02-86f0-1bfd3de3d690-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.968210 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-client-ca" (OuterVolumeSpecName: "client-ca") pod "d00bf964-db4a-449d-ad97-d563b4b70f50" (UID: "d00bf964-db4a-449d-ad97-d563b4b70f50"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.968247 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-config" (OuterVolumeSpecName: "config") pod "d00bf964-db4a-449d-ad97-d563b4b70f50" (UID: "d00bf964-db4a-449d-ad97-d563b4b70f50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.970773 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d00bf964-db4a-449d-ad97-d563b4b70f50-kube-api-access-6g9n9" (OuterVolumeSpecName: "kube-api-access-6g9n9") pod "d00bf964-db4a-449d-ad97-d563b4b70f50" (UID: "d00bf964-db4a-449d-ad97-d563b4b70f50"). InnerVolumeSpecName "kube-api-access-6g9n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:29:02 crc kubenswrapper[4912]: I1203 00:29:02.971696 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d00bf964-db4a-449d-ad97-d563b4b70f50-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d00bf964-db4a-449d-ad97-d563b4b70f50" (UID: "d00bf964-db4a-449d-ad97-d563b4b70f50"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.068777 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.068879 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d00bf964-db4a-449d-ad97-d563b4b70f50-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.068896 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g9n9\" (UniqueName: \"kubernetes.io/projected/d00bf964-db4a-449d-ad97-d563b4b70f50-kube-api-access-6g9n9\") on node \"crc\" DevicePath \"\"" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.068907 4912 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d00bf964-db4a-449d-ad97-d563b4b70f50-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.117351 4912 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.539307 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" event={"ID":"f33e67e7-af60-4c02-86f0-1bfd3de3d690","Type":"ContainerDied","Data":"db2a63aea8194bf1e4ab0a2a919e87cb2c6211ccf184a05791180f7830c2a677"} Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.539365 4912 scope.go:117] "RemoveContainer" containerID="febe37126fabf8ec245b1ec2adf967cc43b2ae15a06c8444cef431b5bac71e6b" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.539392 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-7x9ms" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.543839 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" event={"ID":"d00bf964-db4a-449d-ad97-d563b4b70f50","Type":"ContainerDied","Data":"90e2d8ed356b3c3b61e7b5470014839fd0a78a8c281ab93e36d4e0e89d42b5c7"} Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.543965 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.581891 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7x9ms"] Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.582169 4912 scope.go:117] "RemoveContainer" containerID="a99002b2a1d9a859d66bdc1b157277ad9fe2c114d9fbf1cb5d519cd3eeb2506f" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.590660 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-7x9ms"] Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.606912 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk"] Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.612062 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8hfzk"] Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.727738 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.908511 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll"] Dec 03 00:29:03 crc kubenswrapper[4912]: E1203 00:29:03.909071 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d00bf964-db4a-449d-ad97-d563b4b70f50" containerName="route-controller-manager" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.909094 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d00bf964-db4a-449d-ad97-d563b4b70f50" containerName="route-controller-manager" Dec 03 00:29:03 crc kubenswrapper[4912]: E1203 00:29:03.909117 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f33e67e7-af60-4c02-86f0-1bfd3de3d690" containerName="controller-manager" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.909124 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f33e67e7-af60-4c02-86f0-1bfd3de3d690" containerName="controller-manager" Dec 03 00:29:03 crc kubenswrapper[4912]: E1203 00:29:03.909135 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.909142 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.909233 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d00bf964-db4a-449d-ad97-d563b4b70f50" containerName="route-controller-manager" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.909247 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f33e67e7-af60-4c02-86f0-1bfd3de3d690" containerName="controller-manager" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.909256 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.909684 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.911484 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.911680 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.911891 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.912005 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.912110 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.912287 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.916937 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd"] Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.917812 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.919954 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.920065 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.920114 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.920259 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.922930 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.934457 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.935213 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll"] Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.957624 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 00:29:03 crc kubenswrapper[4912]: I1203 00:29:03.957862 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd"] Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.089881 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-client-ca\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.089942 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-config\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.089961 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-proxy-ca-bundles\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.090016 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-client-ca\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.090035 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4clcs\" (UniqueName: \"kubernetes.io/projected/aeea84f4-8209-4eda-abbd-e90a33dafcb7-kube-api-access-4clcs\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.090061 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aeea84f4-8209-4eda-abbd-e90a33dafcb7-serving-cert\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.090090 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/111e50ae-1c2f-4236-8759-22579b58ca56-serving-cert\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.090151 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fdnd\" (UniqueName: \"kubernetes.io/projected/111e50ae-1c2f-4236-8759-22579b58ca56-kube-api-access-9fdnd\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.090169 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-config\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.190926 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fdnd\" (UniqueName: \"kubernetes.io/projected/111e50ae-1c2f-4236-8759-22579b58ca56-kube-api-access-9fdnd\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.191254 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-config\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.191298 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-client-ca\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.191328 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-config\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.191366 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-proxy-ca-bundles\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.191403 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-client-ca\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.191448 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4clcs\" (UniqueName: \"kubernetes.io/projected/aeea84f4-8209-4eda-abbd-e90a33dafcb7-kube-api-access-4clcs\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.191484 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aeea84f4-8209-4eda-abbd-e90a33dafcb7-serving-cert\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.191511 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/111e50ae-1c2f-4236-8759-22579b58ca56-serving-cert\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.192679 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-client-ca\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.192799 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-proxy-ca-bundles\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.192886 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-config\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.193107 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-config\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.193185 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-client-ca\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.197278 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/111e50ae-1c2f-4236-8759-22579b58ca56-serving-cert\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.197304 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aeea84f4-8209-4eda-abbd-e90a33dafcb7-serving-cert\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.209895 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4clcs\" (UniqueName: \"kubernetes.io/projected/aeea84f4-8209-4eda-abbd-e90a33dafcb7-kube-api-access-4clcs\") pod \"route-controller-manager-685f6d45b6-hz7ll\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.210503 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fdnd\" (UniqueName: \"kubernetes.io/projected/111e50ae-1c2f-4236-8759-22579b58ca56-kube-api-access-9fdnd\") pod \"controller-manager-7bcdf8b5f6-fgzrd\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.255214 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.264604 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.451341 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll"] Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.487971 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd"] Dec 03 00:29:04 crc kubenswrapper[4912]: W1203 00:29:04.500023 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod111e50ae_1c2f_4236_8759_22579b58ca56.slice/crio-bbc43bf45089c75dd5cb5e4cf5ddeb3abf7d60be788ae4f93767465c7d3aee26 WatchSource:0}: Error finding container bbc43bf45089c75dd5cb5e4cf5ddeb3abf7d60be788ae4f93767465c7d3aee26: Status 404 returned error can't find the container with id bbc43bf45089c75dd5cb5e4cf5ddeb3abf7d60be788ae4f93767465c7d3aee26 Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.540723 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.549181 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.559186 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" event={"ID":"111e50ae-1c2f-4236-8759-22579b58ca56","Type":"ContainerStarted","Data":"bbc43bf45089c75dd5cb5e4cf5ddeb3abf7d60be788ae4f93767465c7d3aee26"} Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.562723 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" event={"ID":"aeea84f4-8209-4eda-abbd-e90a33dafcb7","Type":"ContainerStarted","Data":"fc8bc50073fd38860c83d793b021eaf2d7cc1766495f759469dabfb9ecce7336"} Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.580419 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d00bf964-db4a-449d-ad97-d563b4b70f50" path="/var/lib/kubelet/pods/d00bf964-db4a-449d-ad97-d563b4b70f50/volumes" Dec 03 00:29:04 crc kubenswrapper[4912]: I1203 00:29:04.581702 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f33e67e7-af60-4c02-86f0-1bfd3de3d690" path="/var/lib/kubelet/pods/f33e67e7-af60-4c02-86f0-1bfd3de3d690/volumes" Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.299656 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.500687 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.570139 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" event={"ID":"aeea84f4-8209-4eda-abbd-e90a33dafcb7","Type":"ContainerStarted","Data":"acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558"} Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.570403 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.571604 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" event={"ID":"111e50ae-1c2f-4236-8759-22579b58ca56","Type":"ContainerStarted","Data":"1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db"} Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.571820 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.576726 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.577951 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.587986 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" podStartSLOduration=3.5879689089999998 podStartE2EDuration="3.587968909s" podCreationTimestamp="2025-12-03 00:29:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:29:05.587572448 +0000 UTC m=+331.229593018" watchObservedRunningTime="2025-12-03 00:29:05.587968909 +0000 UTC m=+331.229989469" Dec 03 00:29:05 crc kubenswrapper[4912]: I1203 00:29:05.631876 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" podStartSLOduration=3.631859227 podStartE2EDuration="3.631859227s" podCreationTimestamp="2025-12-03 00:29:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:29:05.628862493 +0000 UTC m=+331.270883073" watchObservedRunningTime="2025-12-03 00:29:05.631859227 +0000 UTC m=+331.273879787" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.286824 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nblbp"] Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.288395 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.299687 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nblbp"] Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.423584 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-bound-sa-token\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.423878 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-registry-certificates\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.423954 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-registry-tls\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.424074 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r7qk\" (UniqueName: \"kubernetes.io/projected/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-kube-api-access-4r7qk\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.424633 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.424729 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-trusted-ca\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.424763 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.424871 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.445145 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.526953 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-bound-sa-token\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.527015 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-registry-certificates\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.527037 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-registry-tls\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.527054 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r7qk\" (UniqueName: \"kubernetes.io/projected/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-kube-api-access-4r7qk\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.527082 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.527115 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-trusted-ca\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.527159 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.528571 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-registry-certificates\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.529835 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.532015 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-trusted-ca\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.533640 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-registry-tls\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.533904 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.543884 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r7qk\" (UniqueName: \"kubernetes.io/projected/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-kube-api-access-4r7qk\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.545616 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d-bound-sa-token\") pod \"image-registry-66df7c8f76-nblbp\" (UID: \"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d\") " pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:40 crc kubenswrapper[4912]: I1203 00:29:40.606625 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:41 crc kubenswrapper[4912]: I1203 00:29:41.029412 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nblbp"] Dec 03 00:29:41 crc kubenswrapper[4912]: I1203 00:29:41.785559 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" event={"ID":"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d","Type":"ContainerStarted","Data":"920864f53177310c3d4431c6c9f29b1e25e4ca3a1a8e9e505c098e9737a645e6"} Dec 03 00:29:41 crc kubenswrapper[4912]: I1203 00:29:41.786985 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:29:41 crc kubenswrapper[4912]: I1203 00:29:41.787039 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" event={"ID":"e9c05f8e-c2d2-4d47-aed3-bab8ce15dd8d","Type":"ContainerStarted","Data":"41ea65522688036e6605961b4e81fc407e0385d8823b2aa908e16c1cc6a44c86"} Dec 03 00:29:41 crc kubenswrapper[4912]: I1203 00:29:41.807833 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" podStartSLOduration=1.8078167330000001 podStartE2EDuration="1.807816733s" podCreationTimestamp="2025-12-03 00:29:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:29:41.803368497 +0000 UTC m=+367.445389077" watchObservedRunningTime="2025-12-03 00:29:41.807816733 +0000 UTC m=+367.449837283" Dec 03 00:29:48 crc kubenswrapper[4912]: I1203 00:29:48.077607 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:29:48 crc kubenswrapper[4912]: I1203 00:29:48.078041 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.166819 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554"] Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.168273 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.171722 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.176598 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.210292 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554"] Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.276986 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfec43c-15ce-4a27-b6f2-3d75fe92d634-config-volume\") pod \"collect-profiles-29412030-2f554\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.277044 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6vnb\" (UniqueName: \"kubernetes.io/projected/adfec43c-15ce-4a27-b6f2-3d75fe92d634-kube-api-access-q6vnb\") pod \"collect-profiles-29412030-2f554\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.277194 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfec43c-15ce-4a27-b6f2-3d75fe92d634-secret-volume\") pod \"collect-profiles-29412030-2f554\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.378712 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6vnb\" (UniqueName: \"kubernetes.io/projected/adfec43c-15ce-4a27-b6f2-3d75fe92d634-kube-api-access-q6vnb\") pod \"collect-profiles-29412030-2f554\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.378794 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfec43c-15ce-4a27-b6f2-3d75fe92d634-secret-volume\") pod \"collect-profiles-29412030-2f554\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.378834 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfec43c-15ce-4a27-b6f2-3d75fe92d634-config-volume\") pod \"collect-profiles-29412030-2f554\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.379707 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfec43c-15ce-4a27-b6f2-3d75fe92d634-config-volume\") pod \"collect-profiles-29412030-2f554\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.385175 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfec43c-15ce-4a27-b6f2-3d75fe92d634-secret-volume\") pod \"collect-profiles-29412030-2f554\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.393787 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6vnb\" (UniqueName: \"kubernetes.io/projected/adfec43c-15ce-4a27-b6f2-3d75fe92d634-kube-api-access-q6vnb\") pod \"collect-profiles-29412030-2f554\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.489150 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.689648 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-nblbp" Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.775474 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbppn"] Dec 03 00:30:00 crc kubenswrapper[4912]: I1203 00:30:00.924879 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554"] Dec 03 00:30:01 crc kubenswrapper[4912]: I1203 00:30:01.805321 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd"] Dec 03 00:30:01 crc kubenswrapper[4912]: I1203 00:30:01.805901 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" podUID="111e50ae-1c2f-4236-8759-22579b58ca56" containerName="controller-manager" containerID="cri-o://1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db" gracePeriod=30 Dec 03 00:30:01 crc kubenswrapper[4912]: I1203 00:30:01.817732 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll"] Dec 03 00:30:01 crc kubenswrapper[4912]: I1203 00:30:01.817983 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" podUID="aeea84f4-8209-4eda-abbd-e90a33dafcb7" containerName="route-controller-manager" containerID="cri-o://acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558" gracePeriod=30 Dec 03 00:30:01 crc kubenswrapper[4912]: I1203 00:30:01.914192 4912 generic.go:334] "Generic (PLEG): container finished" podID="adfec43c-15ce-4a27-b6f2-3d75fe92d634" containerID="b9943dc5d01de284185db59710f6de9a9aa6a45df7ba4b4a7bc40169faf6ae47" exitCode=0 Dec 03 00:30:01 crc kubenswrapper[4912]: I1203 00:30:01.914234 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" event={"ID":"adfec43c-15ce-4a27-b6f2-3d75fe92d634","Type":"ContainerDied","Data":"b9943dc5d01de284185db59710f6de9a9aa6a45df7ba4b4a7bc40169faf6ae47"} Dec 03 00:30:01 crc kubenswrapper[4912]: I1203 00:30:01.914260 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" event={"ID":"adfec43c-15ce-4a27-b6f2-3d75fe92d634","Type":"ContainerStarted","Data":"5cee3baacdf725a7b6b776601fb3cb2b753e87eea4a3aeef0d99bf98ba3156ab"} Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.286894 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.295782 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.304400 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-client-ca\") pod \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.305200 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4clcs\" (UniqueName: \"kubernetes.io/projected/aeea84f4-8209-4eda-abbd-e90a33dafcb7-kube-api-access-4clcs\") pod \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.305235 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aeea84f4-8209-4eda-abbd-e90a33dafcb7-serving-cert\") pod \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.305251 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-config\") pod \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\" (UID: \"aeea84f4-8209-4eda-abbd-e90a33dafcb7\") " Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.306154 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-config" (OuterVolumeSpecName: "config") pod "aeea84f4-8209-4eda-abbd-e90a33dafcb7" (UID: "aeea84f4-8209-4eda-abbd-e90a33dafcb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.306709 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-client-ca" (OuterVolumeSpecName: "client-ca") pod "aeea84f4-8209-4eda-abbd-e90a33dafcb7" (UID: "aeea84f4-8209-4eda-abbd-e90a33dafcb7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.311763 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeea84f4-8209-4eda-abbd-e90a33dafcb7-kube-api-access-4clcs" (OuterVolumeSpecName: "kube-api-access-4clcs") pod "aeea84f4-8209-4eda-abbd-e90a33dafcb7" (UID: "aeea84f4-8209-4eda-abbd-e90a33dafcb7"). InnerVolumeSpecName "kube-api-access-4clcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.313173 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeea84f4-8209-4eda-abbd-e90a33dafcb7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "aeea84f4-8209-4eda-abbd-e90a33dafcb7" (UID: "aeea84f4-8209-4eda-abbd-e90a33dafcb7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.406742 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fdnd\" (UniqueName: \"kubernetes.io/projected/111e50ae-1c2f-4236-8759-22579b58ca56-kube-api-access-9fdnd\") pod \"111e50ae-1c2f-4236-8759-22579b58ca56\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.406995 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-config\") pod \"111e50ae-1c2f-4236-8759-22579b58ca56\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.407116 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/111e50ae-1c2f-4236-8759-22579b58ca56-serving-cert\") pod \"111e50ae-1c2f-4236-8759-22579b58ca56\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.407331 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-proxy-ca-bundles\") pod \"111e50ae-1c2f-4236-8759-22579b58ca56\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.407512 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-client-ca\") pod \"111e50ae-1c2f-4236-8759-22579b58ca56\" (UID: \"111e50ae-1c2f-4236-8759-22579b58ca56\") " Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.407763 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-config" (OuterVolumeSpecName: "config") pod "111e50ae-1c2f-4236-8759-22579b58ca56" (UID: "111e50ae-1c2f-4236-8759-22579b58ca56"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.407811 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "111e50ae-1c2f-4236-8759-22579b58ca56" (UID: "111e50ae-1c2f-4236-8759-22579b58ca56"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.407994 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-client-ca" (OuterVolumeSpecName: "client-ca") pod "111e50ae-1c2f-4236-8759-22579b58ca56" (UID: "111e50ae-1c2f-4236-8759-22579b58ca56"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.408523 4912 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.408618 4912 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.408682 4912 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.408740 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4clcs\" (UniqueName: \"kubernetes.io/projected/aeea84f4-8209-4eda-abbd-e90a33dafcb7-kube-api-access-4clcs\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.408799 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/111e50ae-1c2f-4236-8759-22579b58ca56-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.408899 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aeea84f4-8209-4eda-abbd-e90a33dafcb7-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.408963 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeea84f4-8209-4eda-abbd-e90a33dafcb7-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.410045 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/111e50ae-1c2f-4236-8759-22579b58ca56-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "111e50ae-1c2f-4236-8759-22579b58ca56" (UID: "111e50ae-1c2f-4236-8759-22579b58ca56"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.410316 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/111e50ae-1c2f-4236-8759-22579b58ca56-kube-api-access-9fdnd" (OuterVolumeSpecName: "kube-api-access-9fdnd") pod "111e50ae-1c2f-4236-8759-22579b58ca56" (UID: "111e50ae-1c2f-4236-8759-22579b58ca56"). InnerVolumeSpecName "kube-api-access-9fdnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.510502 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fdnd\" (UniqueName: \"kubernetes.io/projected/111e50ae-1c2f-4236-8759-22579b58ca56-kube-api-access-9fdnd\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.510538 4912 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/111e50ae-1c2f-4236-8759-22579b58ca56-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.920932 4912 generic.go:334] "Generic (PLEG): container finished" podID="111e50ae-1c2f-4236-8759-22579b58ca56" containerID="1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db" exitCode=0 Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.920986 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" event={"ID":"111e50ae-1c2f-4236-8759-22579b58ca56","Type":"ContainerDied","Data":"1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db"} Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.921013 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.921036 4912 scope.go:117] "RemoveContainer" containerID="1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.921024 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd" event={"ID":"111e50ae-1c2f-4236-8759-22579b58ca56","Type":"ContainerDied","Data":"bbc43bf45089c75dd5cb5e4cf5ddeb3abf7d60be788ae4f93767465c7d3aee26"} Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.923427 4912 generic.go:334] "Generic (PLEG): container finished" podID="aeea84f4-8209-4eda-abbd-e90a33dafcb7" containerID="acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558" exitCode=0 Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.923464 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.923502 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" event={"ID":"aeea84f4-8209-4eda-abbd-e90a33dafcb7","Type":"ContainerDied","Data":"acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558"} Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.923546 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll" event={"ID":"aeea84f4-8209-4eda-abbd-e90a33dafcb7","Type":"ContainerDied","Data":"fc8bc50073fd38860c83d793b021eaf2d7cc1766495f759469dabfb9ecce7336"} Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.946126 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd"] Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.956884 4912 scope.go:117] "RemoveContainer" containerID="1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.958587 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7bcdf8b5f6-fgzrd"] Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.962715 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll"] Dec 03 00:30:02 crc kubenswrapper[4912]: E1203 00:30:02.963012 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db\": container with ID starting with 1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db not found: ID does not exist" containerID="1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.963048 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db"} err="failed to get container status \"1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db\": rpc error: code = NotFound desc = could not find container \"1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db\": container with ID starting with 1a9f02888f36ff447da58968415895b11b9867d7116827f453157ff0130842db not found: ID does not exist" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.963079 4912 scope.go:117] "RemoveContainer" containerID="acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.965800 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-685f6d45b6-hz7ll"] Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.976792 4912 scope.go:117] "RemoveContainer" containerID="acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558" Dec 03 00:30:02 crc kubenswrapper[4912]: E1203 00:30:02.977188 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558\": container with ID starting with acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558 not found: ID does not exist" containerID="acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558" Dec 03 00:30:02 crc kubenswrapper[4912]: I1203 00:30:02.977231 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558"} err="failed to get container status \"acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558\": rpc error: code = NotFound desc = could not find container \"acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558\": container with ID starting with acf628c6e826d307b7076ff01979a53b8dbf99b7da5a1f194de7b6515a508558 not found: ID does not exist" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.154955 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9"] Dec 03 00:30:03 crc kubenswrapper[4912]: E1203 00:30:03.155292 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="111e50ae-1c2f-4236-8759-22579b58ca56" containerName="controller-manager" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.155312 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="111e50ae-1c2f-4236-8759-22579b58ca56" containerName="controller-manager" Dec 03 00:30:03 crc kubenswrapper[4912]: E1203 00:30:03.155333 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeea84f4-8209-4eda-abbd-e90a33dafcb7" containerName="route-controller-manager" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.155341 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeea84f4-8209-4eda-abbd-e90a33dafcb7" containerName="route-controller-manager" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.155479 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeea84f4-8209-4eda-abbd-e90a33dafcb7" containerName="route-controller-manager" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.155497 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="111e50ae-1c2f-4236-8759-22579b58ca56" containerName="controller-manager" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.155915 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.159020 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6cdd57865d-vt7q2"] Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.159979 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.160644 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.161497 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.161719 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.161721 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.161876 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.161981 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.162292 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.182148 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.182540 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.182561 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.182635 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.182727 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.184859 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.187867 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.194966 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cdd57865d-vt7q2"] Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.199687 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9"] Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.223903 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6vnb\" (UniqueName: \"kubernetes.io/projected/adfec43c-15ce-4a27-b6f2-3d75fe92d634-kube-api-access-q6vnb\") pod \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.224187 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfec43c-15ce-4a27-b6f2-3d75fe92d634-config-volume\") pod \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.224422 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfec43c-15ce-4a27-b6f2-3d75fe92d634-secret-volume\") pod \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\" (UID: \"adfec43c-15ce-4a27-b6f2-3d75fe92d634\") " Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.224835 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b40d4b40-fc53-4555-91a4-4bcf7dd05665-serving-cert\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.225027 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adfec43c-15ce-4a27-b6f2-3d75fe92d634-config-volume" (OuterVolumeSpecName: "config-volume") pod "adfec43c-15ce-4a27-b6f2-3d75fe92d634" (UID: "adfec43c-15ce-4a27-b6f2-3d75fe92d634"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.225042 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a3fffa3-c120-42ce-b909-73425bfcaa64-proxy-ca-bundles\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.225174 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a3fffa3-c120-42ce-b909-73425bfcaa64-config\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.225206 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvm2p\" (UniqueName: \"kubernetes.io/projected/b40d4b40-fc53-4555-91a4-4bcf7dd05665-kube-api-access-lvm2p\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.225239 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a3fffa3-c120-42ce-b909-73425bfcaa64-client-ca\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.225264 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b40d4b40-fc53-4555-91a4-4bcf7dd05665-client-ca\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.225286 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a3fffa3-c120-42ce-b909-73425bfcaa64-serving-cert\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.225330 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9r82\" (UniqueName: \"kubernetes.io/projected/8a3fffa3-c120-42ce-b909-73425bfcaa64-kube-api-access-m9r82\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.225352 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b40d4b40-fc53-4555-91a4-4bcf7dd05665-config\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.228896 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adfec43c-15ce-4a27-b6f2-3d75fe92d634-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "adfec43c-15ce-4a27-b6f2-3d75fe92d634" (UID: "adfec43c-15ce-4a27-b6f2-3d75fe92d634"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.229120 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adfec43c-15ce-4a27-b6f2-3d75fe92d634-kube-api-access-q6vnb" (OuterVolumeSpecName: "kube-api-access-q6vnb") pod "adfec43c-15ce-4a27-b6f2-3d75fe92d634" (UID: "adfec43c-15ce-4a27-b6f2-3d75fe92d634"). InnerVolumeSpecName "kube-api-access-q6vnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.326969 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b40d4b40-fc53-4555-91a4-4bcf7dd05665-serving-cert\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327043 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a3fffa3-c120-42ce-b909-73425bfcaa64-proxy-ca-bundles\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327069 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a3fffa3-c120-42ce-b909-73425bfcaa64-config\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327092 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvm2p\" (UniqueName: \"kubernetes.io/projected/b40d4b40-fc53-4555-91a4-4bcf7dd05665-kube-api-access-lvm2p\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327118 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a3fffa3-c120-42ce-b909-73425bfcaa64-client-ca\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327137 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b40d4b40-fc53-4555-91a4-4bcf7dd05665-client-ca\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327152 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a3fffa3-c120-42ce-b909-73425bfcaa64-serving-cert\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327182 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9r82\" (UniqueName: \"kubernetes.io/projected/8a3fffa3-c120-42ce-b909-73425bfcaa64-kube-api-access-m9r82\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327201 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b40d4b40-fc53-4555-91a4-4bcf7dd05665-config\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327271 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adfec43c-15ce-4a27-b6f2-3d75fe92d634-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327285 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adfec43c-15ce-4a27-b6f2-3d75fe92d634-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.327294 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6vnb\" (UniqueName: \"kubernetes.io/projected/adfec43c-15ce-4a27-b6f2-3d75fe92d634-kube-api-access-q6vnb\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.328493 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b40d4b40-fc53-4555-91a4-4bcf7dd05665-client-ca\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.328600 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b40d4b40-fc53-4555-91a4-4bcf7dd05665-config\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.329237 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a3fffa3-c120-42ce-b909-73425bfcaa64-config\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.330232 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a3fffa3-c120-42ce-b909-73425bfcaa64-client-ca\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.331018 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a3fffa3-c120-42ce-b909-73425bfcaa64-serving-cert\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.332198 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8a3fffa3-c120-42ce-b909-73425bfcaa64-proxy-ca-bundles\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.335074 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b40d4b40-fc53-4555-91a4-4bcf7dd05665-serving-cert\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.345727 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9r82\" (UniqueName: \"kubernetes.io/projected/8a3fffa3-c120-42ce-b909-73425bfcaa64-kube-api-access-m9r82\") pod \"controller-manager-6cdd57865d-vt7q2\" (UID: \"8a3fffa3-c120-42ce-b909-73425bfcaa64\") " pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.346453 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvm2p\" (UniqueName: \"kubernetes.io/projected/b40d4b40-fc53-4555-91a4-4bcf7dd05665-kube-api-access-lvm2p\") pod \"route-controller-manager-565b4f9584-s4jt9\" (UID: \"b40d4b40-fc53-4555-91a4-4bcf7dd05665\") " pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.487559 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.494956 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.689773 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9"] Dec 03 00:30:03 crc kubenswrapper[4912]: W1203 00:30:03.708712 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb40d4b40_fc53_4555_91a4_4bcf7dd05665.slice/crio-9dd1498eeafa6e1a4da739abccdfa733928704d03dd0960f8b60d08380291ad4 WatchSource:0}: Error finding container 9dd1498eeafa6e1a4da739abccdfa733928704d03dd0960f8b60d08380291ad4: Status 404 returned error can't find the container with id 9dd1498eeafa6e1a4da739abccdfa733928704d03dd0960f8b60d08380291ad4 Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.759814 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6cdd57865d-vt7q2"] Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.929271 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" event={"ID":"8a3fffa3-c120-42ce-b909-73425bfcaa64","Type":"ContainerStarted","Data":"bd4e5ad319d85da82db7fa66d1eba692328372de543f855776e9ad8404ed7178"} Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.929714 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.929727 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" event={"ID":"8a3fffa3-c120-42ce-b909-73425bfcaa64","Type":"ContainerStarted","Data":"6aebdfbe9fe75786b129640aeb32f15cae2c41d808b39d8de076144bbe65271e"} Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.931402 4912 patch_prober.go:28] interesting pod/controller-manager-6cdd57865d-vt7q2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" start-of-body= Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.931489 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" podUID="8a3fffa3-c120-42ce-b909-73425bfcaa64" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.932039 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" event={"ID":"b40d4b40-fc53-4555-91a4-4bcf7dd05665","Type":"ContainerStarted","Data":"44d214c5b7da1179bc3d3ccd8554634fd05e6f6fa3caba31436fb70d8d029f78"} Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.932087 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" event={"ID":"b40d4b40-fc53-4555-91a4-4bcf7dd05665","Type":"ContainerStarted","Data":"9dd1498eeafa6e1a4da739abccdfa733928704d03dd0960f8b60d08380291ad4"} Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.932319 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.933294 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" event={"ID":"adfec43c-15ce-4a27-b6f2-3d75fe92d634","Type":"ContainerDied","Data":"5cee3baacdf725a7b6b776601fb3cb2b753e87eea4a3aeef0d99bf98ba3156ab"} Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.933334 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cee3baacdf725a7b6b776601fb3cb2b753e87eea4a3aeef0d99bf98ba3156ab" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.933371 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.954308 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" podStartSLOduration=2.9542864680000003 podStartE2EDuration="2.954286468s" podCreationTimestamp="2025-12-03 00:30:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:30:03.949459553 +0000 UTC m=+389.591480133" watchObservedRunningTime="2025-12-03 00:30:03.954286468 +0000 UTC m=+389.596307028" Dec 03 00:30:03 crc kubenswrapper[4912]: I1203 00:30:03.968388 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" podStartSLOduration=2.968372393 podStartE2EDuration="2.968372393s" podCreationTimestamp="2025-12-03 00:30:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:30:03.964972958 +0000 UTC m=+389.606993528" watchObservedRunningTime="2025-12-03 00:30:03.968372393 +0000 UTC m=+389.610392953" Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.422415 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-565b4f9584-s4jt9" Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.579994 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="111e50ae-1c2f-4236-8759-22579b58ca56" path="/var/lib/kubelet/pods/111e50ae-1c2f-4236-8759-22579b58ca56/volumes" Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.580552 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeea84f4-8209-4eda-abbd-e90a33dafcb7" path="/var/lib/kubelet/pods/aeea84f4-8209-4eda-abbd-e90a33dafcb7/volumes" Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.581132 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kstcn"] Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.581355 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kstcn" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" containerName="registry-server" containerID="cri-o://afd134dceef07e8b04b19944ab1af4c151c50289d445bf245297c42fe9126872" gracePeriod=2 Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.781557 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rf9m5"] Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.781804 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rf9m5" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerName="registry-server" containerID="cri-o://66c92c24baa2245e017a824cc9b5a7d72429c3df4422546d4d78b0ec0065371f" gracePeriod=2 Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.956204 4912 generic.go:334] "Generic (PLEG): container finished" podID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerID="66c92c24baa2245e017a824cc9b5a7d72429c3df4422546d4d78b0ec0065371f" exitCode=0 Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.956598 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf9m5" event={"ID":"179c471f-de9c-4cd6-85ac-b8b3380401ea","Type":"ContainerDied","Data":"66c92c24baa2245e017a824cc9b5a7d72429c3df4422546d4d78b0ec0065371f"} Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.964187 4912 generic.go:334] "Generic (PLEG): container finished" podID="668355d3-1254-4898-a4e9-425b8f480dfd" containerID="afd134dceef07e8b04b19944ab1af4c151c50289d445bf245297c42fe9126872" exitCode=0 Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.965267 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kstcn" event={"ID":"668355d3-1254-4898-a4e9-425b8f480dfd","Type":"ContainerDied","Data":"afd134dceef07e8b04b19944ab1af4c151c50289d445bf245297c42fe9126872"} Dec 03 00:30:04 crc kubenswrapper[4912]: I1203 00:30:04.972077 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6cdd57865d-vt7q2" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.003077 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.056000 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvmqz\" (UniqueName: \"kubernetes.io/projected/668355d3-1254-4898-a4e9-425b8f480dfd-kube-api-access-bvmqz\") pod \"668355d3-1254-4898-a4e9-425b8f480dfd\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.056136 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-utilities\") pod \"668355d3-1254-4898-a4e9-425b8f480dfd\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.056241 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-catalog-content\") pod \"668355d3-1254-4898-a4e9-425b8f480dfd\" (UID: \"668355d3-1254-4898-a4e9-425b8f480dfd\") " Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.058582 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-utilities" (OuterVolumeSpecName: "utilities") pod "668355d3-1254-4898-a4e9-425b8f480dfd" (UID: "668355d3-1254-4898-a4e9-425b8f480dfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.062892 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/668355d3-1254-4898-a4e9-425b8f480dfd-kube-api-access-bvmqz" (OuterVolumeSpecName: "kube-api-access-bvmqz") pod "668355d3-1254-4898-a4e9-425b8f480dfd" (UID: "668355d3-1254-4898-a4e9-425b8f480dfd"). InnerVolumeSpecName "kube-api-access-bvmqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.158246 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvmqz\" (UniqueName: \"kubernetes.io/projected/668355d3-1254-4898-a4e9-425b8f480dfd-kube-api-access-bvmqz\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.158284 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.171654 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.173400 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "668355d3-1254-4898-a4e9-425b8f480dfd" (UID: "668355d3-1254-4898-a4e9-425b8f480dfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.259622 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-catalog-content\") pod \"179c471f-de9c-4cd6-85ac-b8b3380401ea\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.259672 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cmq2\" (UniqueName: \"kubernetes.io/projected/179c471f-de9c-4cd6-85ac-b8b3380401ea-kube-api-access-4cmq2\") pod \"179c471f-de9c-4cd6-85ac-b8b3380401ea\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.259752 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-utilities\") pod \"179c471f-de9c-4cd6-85ac-b8b3380401ea\" (UID: \"179c471f-de9c-4cd6-85ac-b8b3380401ea\") " Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.260031 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668355d3-1254-4898-a4e9-425b8f480dfd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.260579 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-utilities" (OuterVolumeSpecName: "utilities") pod "179c471f-de9c-4cd6-85ac-b8b3380401ea" (UID: "179c471f-de9c-4cd6-85ac-b8b3380401ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.263057 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/179c471f-de9c-4cd6-85ac-b8b3380401ea-kube-api-access-4cmq2" (OuterVolumeSpecName: "kube-api-access-4cmq2") pod "179c471f-de9c-4cd6-85ac-b8b3380401ea" (UID: "179c471f-de9c-4cd6-85ac-b8b3380401ea"). InnerVolumeSpecName "kube-api-access-4cmq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.305371 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "179c471f-de9c-4cd6-85ac-b8b3380401ea" (UID: "179c471f-de9c-4cd6-85ac-b8b3380401ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.361606 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.361651 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cmq2\" (UniqueName: \"kubernetes.io/projected/179c471f-de9c-4cd6-85ac-b8b3380401ea-kube-api-access-4cmq2\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.361668 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179c471f-de9c-4cd6-85ac-b8b3380401ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.973372 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kstcn" event={"ID":"668355d3-1254-4898-a4e9-425b8f480dfd","Type":"ContainerDied","Data":"7ef0721aaa32c2c5ec6ed2dbca3d6294a6854641ee53994cc2ea7162b9652064"} Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.973415 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kstcn" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.973466 4912 scope.go:117] "RemoveContainer" containerID="afd134dceef07e8b04b19944ab1af4c151c50289d445bf245297c42fe9126872" Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.976042 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rf9m5" event={"ID":"179c471f-de9c-4cd6-85ac-b8b3380401ea","Type":"ContainerDied","Data":"f0a4a1486e6ac2cf6da1b6a26ee85eb9bc950027c8ff3819fadd6b51872c36a4"} Dec 03 00:30:05 crc kubenswrapper[4912]: I1203 00:30:05.976217 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rf9m5" Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:05.998907 4912 scope.go:117] "RemoveContainer" containerID="c476afc2810b40d75f6454ff0517094ea36b78cb8c8e87332db8def3f936c30d" Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.030551 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rf9m5"] Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.041290 4912 scope.go:117] "RemoveContainer" containerID="036905eca3af59e5936a201fb04c7100d52f24b312eee0778466c87170c8b666" Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.041599 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rf9m5"] Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.055225 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kstcn"] Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.058007 4912 scope.go:117] "RemoveContainer" containerID="66c92c24baa2245e017a824cc9b5a7d72429c3df4422546d4d78b0ec0065371f" Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.060399 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kstcn"] Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.088127 4912 scope.go:117] "RemoveContainer" containerID="cddddac369d742a7b9a95fa07d4156cfa28dd7e923439d07f752f96b57ef0e8b" Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.110618 4912 scope.go:117] "RemoveContainer" containerID="26678c1c72f5e46cf74cba167b00d69911bf529e770d947d9ea2343c980c22fb" Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.612377 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" path="/var/lib/kubelet/pods/179c471f-de9c-4cd6-85ac-b8b3380401ea/volumes" Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.613650 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" path="/var/lib/kubelet/pods/668355d3-1254-4898-a4e9-425b8f480dfd/volumes" Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.979569 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9p49z"] Dec 03 00:30:06 crc kubenswrapper[4912]: I1203 00:30:06.979812 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9p49z" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerName="registry-server" containerID="cri-o://3cadf8ab231a6629046e59447684578a8c73d7ba4f1a8ba71b93f45cdb736468" gracePeriod=2 Dec 03 00:30:07 crc kubenswrapper[4912]: I1203 00:30:07.190395 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvg59"] Dec 03 00:30:07 crc kubenswrapper[4912]: I1203 00:30:07.190959 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zvg59" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerName="registry-server" containerID="cri-o://155f4fc5456965bfc224b6fe79773d10bc26e0d5edc3260476672fbb6f87a620" gracePeriod=2 Dec 03 00:30:08 crc kubenswrapper[4912]: I1203 00:30:08.999045 4912 generic.go:334] "Generic (PLEG): container finished" podID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerID="3cadf8ab231a6629046e59447684578a8c73d7ba4f1a8ba71b93f45cdb736468" exitCode=0 Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:08.999267 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p49z" event={"ID":"f9edbdcf-7866-4d27-a29f-5426596b00db","Type":"ContainerDied","Data":"3cadf8ab231a6629046e59447684578a8c73d7ba4f1a8ba71b93f45cdb736468"} Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.014283 4912 generic.go:334] "Generic (PLEG): container finished" podID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerID="155f4fc5456965bfc224b6fe79773d10bc26e0d5edc3260476672fbb6f87a620" exitCode=0 Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.014345 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvg59" event={"ID":"616eb06f-63d7-46e2-a8b0-5162a2e159a2","Type":"ContainerDied","Data":"155f4fc5456965bfc224b6fe79773d10bc26e0d5edc3260476672fbb6f87a620"} Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.233967 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.352562 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-utilities\") pod \"f9edbdcf-7866-4d27-a29f-5426596b00db\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.352628 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-catalog-content\") pod \"f9edbdcf-7866-4d27-a29f-5426596b00db\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.352699 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k87nm\" (UniqueName: \"kubernetes.io/projected/f9edbdcf-7866-4d27-a29f-5426596b00db-kube-api-access-k87nm\") pod \"f9edbdcf-7866-4d27-a29f-5426596b00db\" (UID: \"f9edbdcf-7866-4d27-a29f-5426596b00db\") " Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.354008 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-utilities" (OuterVolumeSpecName: "utilities") pod "f9edbdcf-7866-4d27-a29f-5426596b00db" (UID: "f9edbdcf-7866-4d27-a29f-5426596b00db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.359501 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9edbdcf-7866-4d27-a29f-5426596b00db-kube-api-access-k87nm" (OuterVolumeSpecName: "kube-api-access-k87nm") pod "f9edbdcf-7866-4d27-a29f-5426596b00db" (UID: "f9edbdcf-7866-4d27-a29f-5426596b00db"). InnerVolumeSpecName "kube-api-access-k87nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.421537 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9edbdcf-7866-4d27-a29f-5426596b00db" (UID: "f9edbdcf-7866-4d27-a29f-5426596b00db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.451220 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.454259 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.454292 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9edbdcf-7866-4d27-a29f-5426596b00db-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.454302 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k87nm\" (UniqueName: \"kubernetes.io/projected/f9edbdcf-7866-4d27-a29f-5426596b00db-kube-api-access-k87nm\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.555822 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-catalog-content\") pod \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.555937 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-utilities\") pod \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.555987 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmjq4\" (UniqueName: \"kubernetes.io/projected/616eb06f-63d7-46e2-a8b0-5162a2e159a2-kube-api-access-jmjq4\") pod \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\" (UID: \"616eb06f-63d7-46e2-a8b0-5162a2e159a2\") " Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.557718 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-utilities" (OuterVolumeSpecName: "utilities") pod "616eb06f-63d7-46e2-a8b0-5162a2e159a2" (UID: "616eb06f-63d7-46e2-a8b0-5162a2e159a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.559654 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/616eb06f-63d7-46e2-a8b0-5162a2e159a2-kube-api-access-jmjq4" (OuterVolumeSpecName: "kube-api-access-jmjq4") pod "616eb06f-63d7-46e2-a8b0-5162a2e159a2" (UID: "616eb06f-63d7-46e2-a8b0-5162a2e159a2"). InnerVolumeSpecName "kube-api-access-jmjq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.577445 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "616eb06f-63d7-46e2-a8b0-5162a2e159a2" (UID: "616eb06f-63d7-46e2-a8b0-5162a2e159a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.657799 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.657863 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/616eb06f-63d7-46e2-a8b0-5162a2e159a2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:09 crc kubenswrapper[4912]: I1203 00:30:09.657883 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmjq4\" (UniqueName: \"kubernetes.io/projected/616eb06f-63d7-46e2-a8b0-5162a2e159a2-kube-api-access-jmjq4\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.024524 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zvg59" event={"ID":"616eb06f-63d7-46e2-a8b0-5162a2e159a2","Type":"ContainerDied","Data":"0d9771f13a7c78bc19f6f348990bd0ee1125a92973038b84e73fb3050b06454b"} Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.024608 4912 scope.go:117] "RemoveContainer" containerID="155f4fc5456965bfc224b6fe79773d10bc26e0d5edc3260476672fbb6f87a620" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.024839 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zvg59" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.033173 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9p49z" event={"ID":"f9edbdcf-7866-4d27-a29f-5426596b00db","Type":"ContainerDied","Data":"a3c0776f93a2e6f75237a66f274a52a03d145742516242c944f7171270f8e181"} Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.033347 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9p49z" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.050782 4912 scope.go:117] "RemoveContainer" containerID="c964550d2d04de2254d4d9bc261a3a2158b02f72ea6085276f85abbed6123214" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.073853 4912 scope.go:117] "RemoveContainer" containerID="f5849234f6357e0d46b42e30436d7074f9c52e6b126359331636f39d44c266c0" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.083215 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvg59"] Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.087420 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zvg59"] Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.094683 4912 scope.go:117] "RemoveContainer" containerID="3cadf8ab231a6629046e59447684578a8c73d7ba4f1a8ba71b93f45cdb736468" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.104443 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9p49z"] Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.109800 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9p49z"] Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.116478 4912 scope.go:117] "RemoveContainer" containerID="0c3620c3e0096d07944633f6a20a319288ca590a5b6225a0f87a20d58b6f1e62" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.137146 4912 scope.go:117] "RemoveContainer" containerID="1d79119955b934defb3407c59f649b7ae3b9559a476a96e4428fd842dc936cc6" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.577879 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" path="/var/lib/kubelet/pods/616eb06f-63d7-46e2-a8b0-5162a2e159a2/volumes" Dec 03 00:30:10 crc kubenswrapper[4912]: I1203 00:30:10.578756 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" path="/var/lib/kubelet/pods/f9edbdcf-7866-4d27-a29f-5426596b00db/volumes" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.494423 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lqdfv"] Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.497330 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lqdfv" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerName="registry-server" containerID="cri-o://3b10c29ea464f9e8f4875488d5563f3a94f6a6d4957ba00da6587daddd0903b7" gracePeriod=30 Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.504257 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6c6m2"] Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.504533 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6c6m2" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerName="registry-server" containerID="cri-o://9b0c40b3fdd92cd29e475b2cd59c20b9e8b66f5ec527c61bea323572bc5727b7" gracePeriod=30 Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.521933 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgskz"] Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.522188 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" containerID="cri-o://c7c24a61b53aa88d2ec8b06474ee499c238c8210ae197b0ce556db6635edaaaf" gracePeriod=30 Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.528637 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5xn5"] Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.528878 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k5xn5" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerName="registry-server" containerID="cri-o://e61aae9d4b0fd8677c6c6da3d4fe41bdc81f1a8e6d4608d0e371e975cd68d6d6" gracePeriod=30 Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.543949 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-24t69"] Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544258 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerName="extract-utilities" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544283 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerName="extract-utilities" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544295 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adfec43c-15ce-4a27-b6f2-3d75fe92d634" containerName="collect-profiles" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544305 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="adfec43c-15ce-4a27-b6f2-3d75fe92d634" containerName="collect-profiles" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544320 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544330 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544340 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" containerName="extract-utilities" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544348 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" containerName="extract-utilities" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544359 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerName="extract-utilities" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544367 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerName="extract-utilities" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544380 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544389 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544399 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerName="extract-content" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544407 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerName="extract-content" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544420 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerName="extract-content" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544445 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerName="extract-content" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544458 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" containerName="extract-content" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544465 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" containerName="extract-content" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544478 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerName="extract-utilities" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544484 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerName="extract-utilities" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544493 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544499 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544507 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544513 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: E1203 00:30:16.544524 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerName="extract-content" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544530 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerName="extract-content" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544629 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9edbdcf-7866-4d27-a29f-5426596b00db" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544642 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="668355d3-1254-4898-a4e9-425b8f480dfd" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544653 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="179c471f-de9c-4cd6-85ac-b8b3380401ea" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544662 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="616eb06f-63d7-46e2-a8b0-5162a2e159a2" containerName="registry-server" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.544669 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="adfec43c-15ce-4a27-b6f2-3d75fe92d634" containerName="collect-profiles" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.545199 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.550932 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pzmmk"] Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.551283 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pzmmk" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" containerName="registry-server" containerID="cri-o://cc88dacdc2b7e927cfd034a9b956d8d4fd9f64a4e62f003b677106286a4a805f" gracePeriod=30 Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.564700 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-24t69"] Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.662502 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh89s\" (UniqueName: \"kubernetes.io/projected/f8dce103-6de5-4159-a1e3-4ae68c513ee7-kube-api-access-lh89s\") pod \"marketplace-operator-79b997595-24t69\" (UID: \"f8dce103-6de5-4159-a1e3-4ae68c513ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.663043 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8dce103-6de5-4159-a1e3-4ae68c513ee7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-24t69\" (UID: \"f8dce103-6de5-4159-a1e3-4ae68c513ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.663089 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8dce103-6de5-4159-a1e3-4ae68c513ee7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-24t69\" (UID: \"f8dce103-6de5-4159-a1e3-4ae68c513ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.764231 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh89s\" (UniqueName: \"kubernetes.io/projected/f8dce103-6de5-4159-a1e3-4ae68c513ee7-kube-api-access-lh89s\") pod \"marketplace-operator-79b997595-24t69\" (UID: \"f8dce103-6de5-4159-a1e3-4ae68c513ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.764279 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8dce103-6de5-4159-a1e3-4ae68c513ee7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-24t69\" (UID: \"f8dce103-6de5-4159-a1e3-4ae68c513ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.764317 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8dce103-6de5-4159-a1e3-4ae68c513ee7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-24t69\" (UID: \"f8dce103-6de5-4159-a1e3-4ae68c513ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.766189 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8dce103-6de5-4159-a1e3-4ae68c513ee7-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-24t69\" (UID: \"f8dce103-6de5-4159-a1e3-4ae68c513ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.770606 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8dce103-6de5-4159-a1e3-4ae68c513ee7-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-24t69\" (UID: \"f8dce103-6de5-4159-a1e3-4ae68c513ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.785773 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh89s\" (UniqueName: \"kubernetes.io/projected/f8dce103-6de5-4159-a1e3-4ae68c513ee7-kube-api-access-lh89s\") pod \"marketplace-operator-79b997595-24t69\" (UID: \"f8dce103-6de5-4159-a1e3-4ae68c513ee7\") " pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:16 crc kubenswrapper[4912]: I1203 00:30:16.865749 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.079861 4912 generic.go:334] "Generic (PLEG): container finished" podID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerID="e61aae9d4b0fd8677c6c6da3d4fe41bdc81f1a8e6d4608d0e371e975cd68d6d6" exitCode=0 Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.079937 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5xn5" event={"ID":"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01","Type":"ContainerDied","Data":"e61aae9d4b0fd8677c6c6da3d4fe41bdc81f1a8e6d4608d0e371e975cd68d6d6"} Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.082027 4912 generic.go:334] "Generic (PLEG): container finished" podID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerID="3b10c29ea464f9e8f4875488d5563f3a94f6a6d4957ba00da6587daddd0903b7" exitCode=0 Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.082094 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lqdfv" event={"ID":"9e57c9b1-dd22-4cc5-ac09-f010c8030170","Type":"ContainerDied","Data":"3b10c29ea464f9e8f4875488d5563f3a94f6a6d4957ba00da6587daddd0903b7"} Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.083822 4912 generic.go:334] "Generic (PLEG): container finished" podID="c13ea42e-9342-428a-8d6d-4fea05277491" containerID="cc88dacdc2b7e927cfd034a9b956d8d4fd9f64a4e62f003b677106286a4a805f" exitCode=0 Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.083876 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzmmk" event={"ID":"c13ea42e-9342-428a-8d6d-4fea05277491","Type":"ContainerDied","Data":"cc88dacdc2b7e927cfd034a9b956d8d4fd9f64a4e62f003b677106286a4a805f"} Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.085390 4912 generic.go:334] "Generic (PLEG): container finished" podID="c916e93d-66d8-436b-841e-c7a92f823912" containerID="c7c24a61b53aa88d2ec8b06474ee499c238c8210ae197b0ce556db6635edaaaf" exitCode=0 Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.085553 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" event={"ID":"c916e93d-66d8-436b-841e-c7a92f823912","Type":"ContainerDied","Data":"c7c24a61b53aa88d2ec8b06474ee499c238c8210ae197b0ce556db6635edaaaf"} Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.085675 4912 scope.go:117] "RemoveContainer" containerID="acb670445e0d86c859c31d49fe0734cd7907ebf6d9586c532ee7efaea4e0c711" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.088812 4912 generic.go:334] "Generic (PLEG): container finished" podID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerID="9b0c40b3fdd92cd29e475b2cd59c20b9e8b66f5ec527c61bea323572bc5727b7" exitCode=0 Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.088890 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6c6m2" event={"ID":"da8155c3-0290-44f3-b1da-e95d2dcee000","Type":"ContainerDied","Data":"9b0c40b3fdd92cd29e475b2cd59c20b9e8b66f5ec527c61bea323572bc5727b7"} Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.130739 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.270737 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4szh6\" (UniqueName: \"kubernetes.io/projected/9e57c9b1-dd22-4cc5-ac09-f010c8030170-kube-api-access-4szh6\") pod \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.270856 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-utilities\") pod \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.270891 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-catalog-content\") pod \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\" (UID: \"9e57c9b1-dd22-4cc5-ac09-f010c8030170\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.272385 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-utilities" (OuterVolumeSpecName: "utilities") pod "9e57c9b1-dd22-4cc5-ac09-f010c8030170" (UID: "9e57c9b1-dd22-4cc5-ac09-f010c8030170"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.276274 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e57c9b1-dd22-4cc5-ac09-f010c8030170-kube-api-access-4szh6" (OuterVolumeSpecName: "kube-api-access-4szh6") pod "9e57c9b1-dd22-4cc5-ac09-f010c8030170" (UID: "9e57c9b1-dd22-4cc5-ac09-f010c8030170"). InnerVolumeSpecName "kube-api-access-4szh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.301788 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.344352 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e57c9b1-dd22-4cc5-ac09-f010c8030170" (UID: "9e57c9b1-dd22-4cc5-ac09-f010c8030170"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.369703 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.371807 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7gvq\" (UniqueName: \"kubernetes.io/projected/da8155c3-0290-44f3-b1da-e95d2dcee000-kube-api-access-b7gvq\") pod \"da8155c3-0290-44f3-b1da-e95d2dcee000\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.371862 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-utilities\") pod \"da8155c3-0290-44f3-b1da-e95d2dcee000\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.371925 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-catalog-content\") pod \"da8155c3-0290-44f3-b1da-e95d2dcee000\" (UID: \"da8155c3-0290-44f3-b1da-e95d2dcee000\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.372211 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.372226 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e57c9b1-dd22-4cc5-ac09-f010c8030170-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.372237 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4szh6\" (UniqueName: \"kubernetes.io/projected/9e57c9b1-dd22-4cc5-ac09-f010c8030170-kube-api-access-4szh6\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.373992 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-utilities" (OuterVolumeSpecName: "utilities") pod "da8155c3-0290-44f3-b1da-e95d2dcee000" (UID: "da8155c3-0290-44f3-b1da-e95d2dcee000"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.374681 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.380261 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da8155c3-0290-44f3-b1da-e95d2dcee000-kube-api-access-b7gvq" (OuterVolumeSpecName: "kube-api-access-b7gvq") pod "da8155c3-0290-44f3-b1da-e95d2dcee000" (UID: "da8155c3-0290-44f3-b1da-e95d2dcee000"). InnerVolumeSpecName "kube-api-access-b7gvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.381023 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.399442 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-24t69"] Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.427856 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da8155c3-0290-44f3-b1da-e95d2dcee000" (UID: "da8155c3-0290-44f3-b1da-e95d2dcee000"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.473381 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-catalog-content\") pod \"c13ea42e-9342-428a-8d6d-4fea05277491\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.473449 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-utilities\") pod \"c13ea42e-9342-428a-8d6d-4fea05277491\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.473491 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4kq4\" (UniqueName: \"kubernetes.io/projected/c13ea42e-9342-428a-8d6d-4fea05277491-kube-api-access-c4kq4\") pod \"c13ea42e-9342-428a-8d6d-4fea05277491\" (UID: \"c13ea42e-9342-428a-8d6d-4fea05277491\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.473523 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-trusted-ca\") pod \"c916e93d-66d8-436b-841e-c7a92f823912\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.473579 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcqx5\" (UniqueName: \"kubernetes.io/projected/c916e93d-66d8-436b-841e-c7a92f823912-kube-api-access-hcqx5\") pod \"c916e93d-66d8-436b-841e-c7a92f823912\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.473607 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6vhn\" (UniqueName: \"kubernetes.io/projected/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-kube-api-access-l6vhn\") pod \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.473678 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-utilities\") pod \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.473724 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-operator-metrics\") pod \"c916e93d-66d8-436b-841e-c7a92f823912\" (UID: \"c916e93d-66d8-436b-841e-c7a92f823912\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.473768 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-catalog-content\") pod \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\" (UID: \"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01\") " Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.474076 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7gvq\" (UniqueName: \"kubernetes.io/projected/da8155c3-0290-44f3-b1da-e95d2dcee000-kube-api-access-b7gvq\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.474093 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.474105 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da8155c3-0290-44f3-b1da-e95d2dcee000-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.474658 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-utilities" (OuterVolumeSpecName: "utilities") pod "c13ea42e-9342-428a-8d6d-4fea05277491" (UID: "c13ea42e-9342-428a-8d6d-4fea05277491"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.474671 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-utilities" (OuterVolumeSpecName: "utilities") pod "0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" (UID: "0c0811e2-2b5f-4dd9-9b2f-6039a3297e01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.475207 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c916e93d-66d8-436b-841e-c7a92f823912" (UID: "c916e93d-66d8-436b-841e-c7a92f823912"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.480384 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c916e93d-66d8-436b-841e-c7a92f823912" (UID: "c916e93d-66d8-436b-841e-c7a92f823912"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.480692 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c916e93d-66d8-436b-841e-c7a92f823912-kube-api-access-hcqx5" (OuterVolumeSpecName: "kube-api-access-hcqx5") pod "c916e93d-66d8-436b-841e-c7a92f823912" (UID: "c916e93d-66d8-436b-841e-c7a92f823912"). InnerVolumeSpecName "kube-api-access-hcqx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.484365 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-kube-api-access-l6vhn" (OuterVolumeSpecName: "kube-api-access-l6vhn") pod "0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" (UID: "0c0811e2-2b5f-4dd9-9b2f-6039a3297e01"). InnerVolumeSpecName "kube-api-access-l6vhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.484388 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c13ea42e-9342-428a-8d6d-4fea05277491-kube-api-access-c4kq4" (OuterVolumeSpecName: "kube-api-access-c4kq4") pod "c13ea42e-9342-428a-8d6d-4fea05277491" (UID: "c13ea42e-9342-428a-8d6d-4fea05277491"). InnerVolumeSpecName "kube-api-access-c4kq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.493256 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" (UID: "0c0811e2-2b5f-4dd9-9b2f-6039a3297e01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.575785 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcqx5\" (UniqueName: \"kubernetes.io/projected/c916e93d-66d8-436b-841e-c7a92f823912-kube-api-access-hcqx5\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.575826 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6vhn\" (UniqueName: \"kubernetes.io/projected/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-kube-api-access-l6vhn\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.575840 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.575856 4912 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.575871 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.575882 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.575894 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4kq4\" (UniqueName: \"kubernetes.io/projected/c13ea42e-9342-428a-8d6d-4fea05277491-kube-api-access-c4kq4\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.575906 4912 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c916e93d-66d8-436b-841e-c7a92f823912-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.577313 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c13ea42e-9342-428a-8d6d-4fea05277491" (UID: "c13ea42e-9342-428a-8d6d-4fea05277491"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:17 crc kubenswrapper[4912]: I1203 00:30:17.676817 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13ea42e-9342-428a-8d6d-4fea05277491-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.077490 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.077544 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.098101 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k5xn5" event={"ID":"0c0811e2-2b5f-4dd9-9b2f-6039a3297e01","Type":"ContainerDied","Data":"2893a29b6e768df7495f4bdf462540c54e545ae2e36465093946dd778202731b"} Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.098161 4912 scope.go:117] "RemoveContainer" containerID="e61aae9d4b0fd8677c6c6da3d4fe41bdc81f1a8e6d4608d0e371e975cd68d6d6" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.098164 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k5xn5" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.101291 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lqdfv" event={"ID":"9e57c9b1-dd22-4cc5-ac09-f010c8030170","Type":"ContainerDied","Data":"c5b02ccb545bedc42f4b6a55418471016e9e43f1ed38a7fe468632080ec3a994"} Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.101319 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lqdfv" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.103066 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-24t69" event={"ID":"f8dce103-6de5-4159-a1e3-4ae68c513ee7","Type":"ContainerStarted","Data":"d447a67e09f7bf13f53b7b453de369f85284ccd09842121b2092a7e7f45235be"} Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.103101 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-24t69" event={"ID":"f8dce103-6de5-4159-a1e3-4ae68c513ee7","Type":"ContainerStarted","Data":"f327a7e079596bce74949b00b21485a7713233dcabef9b7f8ac087ed90fc13ee"} Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.103459 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.105212 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzmmk" event={"ID":"c13ea42e-9342-428a-8d6d-4fea05277491","Type":"ContainerDied","Data":"aa927155ec38e791f42077e456b8cdc07b7d19b4e6b819af4e801d1aa8af01de"} Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.105351 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzmmk" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.106781 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" event={"ID":"c916e93d-66d8-436b-841e-c7a92f823912","Type":"ContainerDied","Data":"135405967aa9740987410dff7772d719cef83a444ba490724d8d97d7f0a93cdd"} Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.106850 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgskz" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.108479 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-24t69" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.113776 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6c6m2" event={"ID":"da8155c3-0290-44f3-b1da-e95d2dcee000","Type":"ContainerDied","Data":"ea5f298bcce23ad702183f6170f2ae9fbc2f1ab4e74aa14e400655a6880ff0e0"} Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.113898 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6c6m2" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.116000 4912 scope.go:117] "RemoveContainer" containerID="b439f3237f472dac6b8c966c73d913ddc9319800bc2f97c58c1cf7356c25eec3" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.128288 4912 scope.go:117] "RemoveContainer" containerID="1fc0ea09da6279922d43a9a483e82a2d947634813381f7aaf1c25aced86081d6" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.164254 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-24t69" podStartSLOduration=2.164239738 podStartE2EDuration="2.164239738s" podCreationTimestamp="2025-12-03 00:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:30:18.143336151 +0000 UTC m=+403.785356741" watchObservedRunningTime="2025-12-03 00:30:18.164239738 +0000 UTC m=+403.806260298" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.168780 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lqdfv"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.179913 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lqdfv"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.186041 4912 scope.go:117] "RemoveContainer" containerID="3b10c29ea464f9e8f4875488d5563f3a94f6a6d4957ba00da6587daddd0903b7" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.188561 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5xn5"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.202543 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k5xn5"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.218173 4912 scope.go:117] "RemoveContainer" containerID="bfa33806ce33d2f56fb7688c916be7c10a27b158c1a04b671788b58a2615eca8" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.234395 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgskz"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.254730 4912 scope.go:117] "RemoveContainer" containerID="c712ff215cb58f2d2b0e28e6f043db095cb833bf12ca25237e9b5ee8a2235fe1" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.256904 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgskz"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.262561 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pzmmk"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.267444 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pzmmk"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.272952 4912 scope.go:117] "RemoveContainer" containerID="cc88dacdc2b7e927cfd034a9b956d8d4fd9f64a4e62f003b677106286a4a805f" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.282016 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6c6m2"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.286325 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6c6m2"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.305690 4912 scope.go:117] "RemoveContainer" containerID="530ef7e8325f79cbe06558aae67e3454e4cdb54ba98cdddf5542475d922c64a7" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.321966 4912 scope.go:117] "RemoveContainer" containerID="b3cbd098a44ed8f847624dd7f5446c941b2cbd63359696f25e19b016152a5085" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.336789 4912 scope.go:117] "RemoveContainer" containerID="c7c24a61b53aa88d2ec8b06474ee499c238c8210ae197b0ce556db6635edaaaf" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.349293 4912 scope.go:117] "RemoveContainer" containerID="9b0c40b3fdd92cd29e475b2cd59c20b9e8b66f5ec527c61bea323572bc5727b7" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.361016 4912 scope.go:117] "RemoveContainer" containerID="6ce1ac9cbf83038a89ef1d372728d58bf30356a3a0cb4e615b5082b7f828d5bd" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.374076 4912 scope.go:117] "RemoveContainer" containerID="5bbfd84cd7b31f524119d50728a4e55a185f6d46162d8c1c8e2f78d4f369ba99" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.587109 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" path="/var/lib/kubelet/pods/0c0811e2-2b5f-4dd9-9b2f-6039a3297e01/volumes" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.587936 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" path="/var/lib/kubelet/pods/9e57c9b1-dd22-4cc5-ac09-f010c8030170/volumes" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.588527 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" path="/var/lib/kubelet/pods/c13ea42e-9342-428a-8d6d-4fea05277491/volumes" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.589620 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c916e93d-66d8-436b-841e-c7a92f823912" path="/var/lib/kubelet/pods/c916e93d-66d8-436b-841e-c7a92f823912/volumes" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.590081 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" path="/var/lib/kubelet/pods/da8155c3-0290-44f3-b1da-e95d2dcee000/volumes" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.710818 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jfl9b"] Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711114 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerName="extract-content" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711136 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerName="extract-content" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711153 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711164 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711178 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerName="extract-content" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711188 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerName="extract-content" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711201 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711210 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711220 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerName="extract-content" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711228 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerName="extract-content" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711241 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerName="extract-utilities" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711250 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerName="extract-utilities" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711265 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711275 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711288 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerName="extract-utilities" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711296 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerName="extract-utilities" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711308 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711316 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711326 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711337 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711349 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerName="extract-utilities" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711358 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerName="extract-utilities" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711367 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" containerName="extract-content" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711375 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" containerName="extract-content" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711387 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" containerName="extract-utilities" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711396 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" containerName="extract-utilities" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711568 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e57c9b1-dd22-4cc5-ac09-f010c8030170" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711596 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711610 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="da8155c3-0290-44f3-b1da-e95d2dcee000" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711622 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c13ea42e-9342-428a-8d6d-4fea05277491" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711640 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c0811e2-2b5f-4dd9-9b2f-6039a3297e01" containerName="registry-server" Dec 03 00:30:18 crc kubenswrapper[4912]: E1203 00:30:18.711763 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711774 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.711890 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c916e93d-66d8-436b-841e-c7a92f823912" containerName="marketplace-operator" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.712670 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.718874 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.720992 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jfl9b"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.790651 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d18182ed-8455-4575-aa3d-99fa4e26f604-catalog-content\") pod \"certified-operators-jfl9b\" (UID: \"d18182ed-8455-4575-aa3d-99fa4e26f604\") " pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.790794 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d18182ed-8455-4575-aa3d-99fa4e26f604-utilities\") pod \"certified-operators-jfl9b\" (UID: \"d18182ed-8455-4575-aa3d-99fa4e26f604\") " pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.790839 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cqzb\" (UniqueName: \"kubernetes.io/projected/d18182ed-8455-4575-aa3d-99fa4e26f604-kube-api-access-6cqzb\") pod \"certified-operators-jfl9b\" (UID: \"d18182ed-8455-4575-aa3d-99fa4e26f604\") " pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.892631 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d18182ed-8455-4575-aa3d-99fa4e26f604-utilities\") pod \"certified-operators-jfl9b\" (UID: \"d18182ed-8455-4575-aa3d-99fa4e26f604\") " pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.892689 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cqzb\" (UniqueName: \"kubernetes.io/projected/d18182ed-8455-4575-aa3d-99fa4e26f604-kube-api-access-6cqzb\") pod \"certified-operators-jfl9b\" (UID: \"d18182ed-8455-4575-aa3d-99fa4e26f604\") " pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.892715 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d18182ed-8455-4575-aa3d-99fa4e26f604-catalog-content\") pod \"certified-operators-jfl9b\" (UID: \"d18182ed-8455-4575-aa3d-99fa4e26f604\") " pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.893269 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d18182ed-8455-4575-aa3d-99fa4e26f604-catalog-content\") pod \"certified-operators-jfl9b\" (UID: \"d18182ed-8455-4575-aa3d-99fa4e26f604\") " pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.893373 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d18182ed-8455-4575-aa3d-99fa4e26f604-utilities\") pod \"certified-operators-jfl9b\" (UID: \"d18182ed-8455-4575-aa3d-99fa4e26f604\") " pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.918614 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cqzb\" (UniqueName: \"kubernetes.io/projected/d18182ed-8455-4575-aa3d-99fa4e26f604-kube-api-access-6cqzb\") pod \"certified-operators-jfl9b\" (UID: \"d18182ed-8455-4575-aa3d-99fa4e26f604\") " pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.919887 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qqv59"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.921116 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.924547 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.929115 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqv59"] Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.993788 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaca572-9fa6-4c8f-8563-0609714513b0-utilities\") pod \"redhat-marketplace-qqv59\" (UID: \"5aaca572-9fa6-4c8f-8563-0609714513b0\") " pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.993871 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42b27\" (UniqueName: \"kubernetes.io/projected/5aaca572-9fa6-4c8f-8563-0609714513b0-kube-api-access-42b27\") pod \"redhat-marketplace-qqv59\" (UID: \"5aaca572-9fa6-4c8f-8563-0609714513b0\") " pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:18 crc kubenswrapper[4912]: I1203 00:30:18.993944 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaca572-9fa6-4c8f-8563-0609714513b0-catalog-content\") pod \"redhat-marketplace-qqv59\" (UID: \"5aaca572-9fa6-4c8f-8563-0609714513b0\") " pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.041979 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.095521 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaca572-9fa6-4c8f-8563-0609714513b0-catalog-content\") pod \"redhat-marketplace-qqv59\" (UID: \"5aaca572-9fa6-4c8f-8563-0609714513b0\") " pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.095598 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaca572-9fa6-4c8f-8563-0609714513b0-utilities\") pod \"redhat-marketplace-qqv59\" (UID: \"5aaca572-9fa6-4c8f-8563-0609714513b0\") " pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.095661 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42b27\" (UniqueName: \"kubernetes.io/projected/5aaca572-9fa6-4c8f-8563-0609714513b0-kube-api-access-42b27\") pod \"redhat-marketplace-qqv59\" (UID: \"5aaca572-9fa6-4c8f-8563-0609714513b0\") " pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.096357 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5aaca572-9fa6-4c8f-8563-0609714513b0-utilities\") pod \"redhat-marketplace-qqv59\" (UID: \"5aaca572-9fa6-4c8f-8563-0609714513b0\") " pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.096367 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5aaca572-9fa6-4c8f-8563-0609714513b0-catalog-content\") pod \"redhat-marketplace-qqv59\" (UID: \"5aaca572-9fa6-4c8f-8563-0609714513b0\") " pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.113487 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42b27\" (UniqueName: \"kubernetes.io/projected/5aaca572-9fa6-4c8f-8563-0609714513b0-kube-api-access-42b27\") pod \"redhat-marketplace-qqv59\" (UID: \"5aaca572-9fa6-4c8f-8563-0609714513b0\") " pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.251801 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.463890 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jfl9b"] Dec 03 00:30:19 crc kubenswrapper[4912]: W1203 00:30:19.471392 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd18182ed_8455_4575_aa3d_99fa4e26f604.slice/crio-7828efd28bb18f3e31543c9f8db7c4460b45fa7790896e80ce79bbeac3b51ecd WatchSource:0}: Error finding container 7828efd28bb18f3e31543c9f8db7c4460b45fa7790896e80ce79bbeac3b51ecd: Status 404 returned error can't find the container with id 7828efd28bb18f3e31543c9f8db7c4460b45fa7790896e80ce79bbeac3b51ecd Dec 03 00:30:19 crc kubenswrapper[4912]: I1203 00:30:19.634270 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqv59"] Dec 03 00:30:19 crc kubenswrapper[4912]: W1203 00:30:19.686466 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5aaca572_9fa6_4c8f_8563_0609714513b0.slice/crio-1e5d7d58c706c3e3de52d0ab1629627d867a17a5484a4ada2c9cd4a6a3c58e3d WatchSource:0}: Error finding container 1e5d7d58c706c3e3de52d0ab1629627d867a17a5484a4ada2c9cd4a6a3c58e3d: Status 404 returned error can't find the container with id 1e5d7d58c706c3e3de52d0ab1629627d867a17a5484a4ada2c9cd4a6a3c58e3d Dec 03 00:30:20 crc kubenswrapper[4912]: I1203 00:30:20.150296 4912 generic.go:334] "Generic (PLEG): container finished" podID="5aaca572-9fa6-4c8f-8563-0609714513b0" containerID="b0386a46c78a200ddc1431031e03911a391059520fb77811f8a48457d512772a" exitCode=0 Dec 03 00:30:20 crc kubenswrapper[4912]: I1203 00:30:20.150340 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqv59" event={"ID":"5aaca572-9fa6-4c8f-8563-0609714513b0","Type":"ContainerDied","Data":"b0386a46c78a200ddc1431031e03911a391059520fb77811f8a48457d512772a"} Dec 03 00:30:20 crc kubenswrapper[4912]: I1203 00:30:20.150377 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqv59" event={"ID":"5aaca572-9fa6-4c8f-8563-0609714513b0","Type":"ContainerStarted","Data":"1e5d7d58c706c3e3de52d0ab1629627d867a17a5484a4ada2c9cd4a6a3c58e3d"} Dec 03 00:30:20 crc kubenswrapper[4912]: I1203 00:30:20.152771 4912 generic.go:334] "Generic (PLEG): container finished" podID="d18182ed-8455-4575-aa3d-99fa4e26f604" containerID="4ca07dd38788f0809377b401f0fd635570e20ee08565b1412a5dfa043019e1d8" exitCode=0 Dec 03 00:30:20 crc kubenswrapper[4912]: I1203 00:30:20.153557 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfl9b" event={"ID":"d18182ed-8455-4575-aa3d-99fa4e26f604","Type":"ContainerDied","Data":"4ca07dd38788f0809377b401f0fd635570e20ee08565b1412a5dfa043019e1d8"} Dec 03 00:30:20 crc kubenswrapper[4912]: I1203 00:30:20.153581 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfl9b" event={"ID":"d18182ed-8455-4575-aa3d-99fa4e26f604","Type":"ContainerStarted","Data":"7828efd28bb18f3e31543c9f8db7c4460b45fa7790896e80ce79bbeac3b51ecd"} Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.112863 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bkzm2"] Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.114167 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.116781 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.124515 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkzm2"] Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.160298 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqv59" event={"ID":"5aaca572-9fa6-4c8f-8563-0609714513b0","Type":"ContainerStarted","Data":"278ed41ed70fbcc252f48b0abe23043f6bac161da539e5884cc15ecbb6a47521"} Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.162877 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfl9b" event={"ID":"d18182ed-8455-4575-aa3d-99fa4e26f604","Type":"ContainerStarted","Data":"bb04b5878fff6f9aa9f171c35158c9a3c57fd6038958e935ec1d9358c8b353c3"} Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.222136 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e9fe039-8cbd-4a92-9b85-4a17818ade89-utilities\") pod \"community-operators-bkzm2\" (UID: \"8e9fe039-8cbd-4a92-9b85-4a17818ade89\") " pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.222382 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e9fe039-8cbd-4a92-9b85-4a17818ade89-catalog-content\") pod \"community-operators-bkzm2\" (UID: \"8e9fe039-8cbd-4a92-9b85-4a17818ade89\") " pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.222419 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdfww\" (UniqueName: \"kubernetes.io/projected/8e9fe039-8cbd-4a92-9b85-4a17818ade89-kube-api-access-mdfww\") pod \"community-operators-bkzm2\" (UID: \"8e9fe039-8cbd-4a92-9b85-4a17818ade89\") " pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.309575 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5xg4j"] Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.310929 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.316106 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.323388 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e9fe039-8cbd-4a92-9b85-4a17818ade89-utilities\") pod \"community-operators-bkzm2\" (UID: \"8e9fe039-8cbd-4a92-9b85-4a17818ade89\") " pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.323421 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e9fe039-8cbd-4a92-9b85-4a17818ade89-catalog-content\") pod \"community-operators-bkzm2\" (UID: \"8e9fe039-8cbd-4a92-9b85-4a17818ade89\") " pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.323474 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdfww\" (UniqueName: \"kubernetes.io/projected/8e9fe039-8cbd-4a92-9b85-4a17818ade89-kube-api-access-mdfww\") pod \"community-operators-bkzm2\" (UID: \"8e9fe039-8cbd-4a92-9b85-4a17818ade89\") " pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.323723 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5xg4j"] Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.324582 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e9fe039-8cbd-4a92-9b85-4a17818ade89-utilities\") pod \"community-operators-bkzm2\" (UID: \"8e9fe039-8cbd-4a92-9b85-4a17818ade89\") " pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.325558 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e9fe039-8cbd-4a92-9b85-4a17818ade89-catalog-content\") pod \"community-operators-bkzm2\" (UID: \"8e9fe039-8cbd-4a92-9b85-4a17818ade89\") " pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.359589 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdfww\" (UniqueName: \"kubernetes.io/projected/8e9fe039-8cbd-4a92-9b85-4a17818ade89-kube-api-access-mdfww\") pod \"community-operators-bkzm2\" (UID: \"8e9fe039-8cbd-4a92-9b85-4a17818ade89\") " pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.425528 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-catalog-content\") pod \"redhat-operators-5xg4j\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.425623 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t89wf\" (UniqueName: \"kubernetes.io/projected/770b4a36-b3be-4828-9e41-3fb24fa639f5-kube-api-access-t89wf\") pod \"redhat-operators-5xg4j\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.425805 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-utilities\") pod \"redhat-operators-5xg4j\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.432094 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.527841 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t89wf\" (UniqueName: \"kubernetes.io/projected/770b4a36-b3be-4828-9e41-3fb24fa639f5-kube-api-access-t89wf\") pod \"redhat-operators-5xg4j\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.527904 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-utilities\") pod \"redhat-operators-5xg4j\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.527962 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-catalog-content\") pod \"redhat-operators-5xg4j\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.528669 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-utilities\") pod \"redhat-operators-5xg4j\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.528922 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-catalog-content\") pod \"redhat-operators-5xg4j\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.547228 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t89wf\" (UniqueName: \"kubernetes.io/projected/770b4a36-b3be-4828-9e41-3fb24fa639f5-kube-api-access-t89wf\") pod \"redhat-operators-5xg4j\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.635857 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:21 crc kubenswrapper[4912]: I1203 00:30:21.843673 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkzm2"] Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.080553 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5xg4j"] Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.173195 4912 generic.go:334] "Generic (PLEG): container finished" podID="5aaca572-9fa6-4c8f-8563-0609714513b0" containerID="278ed41ed70fbcc252f48b0abe23043f6bac161da539e5884cc15ecbb6a47521" exitCode=0 Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.173315 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqv59" event={"ID":"5aaca572-9fa6-4c8f-8563-0609714513b0","Type":"ContainerDied","Data":"278ed41ed70fbcc252f48b0abe23043f6bac161da539e5884cc15ecbb6a47521"} Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.183166 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xg4j" event={"ID":"770b4a36-b3be-4828-9e41-3fb24fa639f5","Type":"ContainerStarted","Data":"d55ecddf073fc89f64f801b2e84d00605670b25fb66a9db4ef9a9ea48dd94f9c"} Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.186567 4912 generic.go:334] "Generic (PLEG): container finished" podID="d18182ed-8455-4575-aa3d-99fa4e26f604" containerID="bb04b5878fff6f9aa9f171c35158c9a3c57fd6038958e935ec1d9358c8b353c3" exitCode=0 Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.186655 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfl9b" event={"ID":"d18182ed-8455-4575-aa3d-99fa4e26f604","Type":"ContainerDied","Data":"bb04b5878fff6f9aa9f171c35158c9a3c57fd6038958e935ec1d9358c8b353c3"} Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.186697 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jfl9b" event={"ID":"d18182ed-8455-4575-aa3d-99fa4e26f604","Type":"ContainerStarted","Data":"362660c79a730defd93bcf6d32c680331acbf0d25186c47e4a27c8f7c688d010"} Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.188284 4912 generic.go:334] "Generic (PLEG): container finished" podID="8e9fe039-8cbd-4a92-9b85-4a17818ade89" containerID="635b2b07e6e434ed75321ff17e78996eeb1b87886ab902c9639e48ebf50da4ea" exitCode=0 Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.188312 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkzm2" event={"ID":"8e9fe039-8cbd-4a92-9b85-4a17818ade89","Type":"ContainerDied","Data":"635b2b07e6e434ed75321ff17e78996eeb1b87886ab902c9639e48ebf50da4ea"} Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.188906 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkzm2" event={"ID":"8e9fe039-8cbd-4a92-9b85-4a17818ade89","Type":"ContainerStarted","Data":"e2c7a354c2c9e784e1a402b951d3362ed61aefc652143a0d4c54ddb43f33afd3"} Dec 03 00:30:22 crc kubenswrapper[4912]: I1203 00:30:22.232629 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jfl9b" podStartSLOduration=2.530287779 podStartE2EDuration="4.232604541s" podCreationTimestamp="2025-12-03 00:30:18 +0000 UTC" firstStartedPulling="2025-12-03 00:30:20.154360715 +0000 UTC m=+405.796381275" lastFinishedPulling="2025-12-03 00:30:21.856677477 +0000 UTC m=+407.498698037" observedRunningTime="2025-12-03 00:30:22.229869647 +0000 UTC m=+407.871890207" watchObservedRunningTime="2025-12-03 00:30:22.232604541 +0000 UTC m=+407.874625101" Dec 03 00:30:23 crc kubenswrapper[4912]: I1203 00:30:23.196477 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkzm2" event={"ID":"8e9fe039-8cbd-4a92-9b85-4a17818ade89","Type":"ContainerStarted","Data":"771573c7c8d0064fd200f0bc4af3610ac9bb5a0e571441de743613ae2cbcc784"} Dec 03 00:30:23 crc kubenswrapper[4912]: I1203 00:30:23.199524 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqv59" event={"ID":"5aaca572-9fa6-4c8f-8563-0609714513b0","Type":"ContainerStarted","Data":"64cca4390bc985fb32444ab27afb892cf05680cb354da3e6a594f35e51e641e1"} Dec 03 00:30:23 crc kubenswrapper[4912]: I1203 00:30:23.204136 4912 generic.go:334] "Generic (PLEG): container finished" podID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerID="189976446c8dec6c87e6f44369faa6313c53cbda2d104c28e46f9cffcaf6e0fe" exitCode=0 Dec 03 00:30:23 crc kubenswrapper[4912]: I1203 00:30:23.204871 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xg4j" event={"ID":"770b4a36-b3be-4828-9e41-3fb24fa639f5","Type":"ContainerDied","Data":"189976446c8dec6c87e6f44369faa6313c53cbda2d104c28e46f9cffcaf6e0fe"} Dec 03 00:30:23 crc kubenswrapper[4912]: I1203 00:30:23.268274 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qqv59" podStartSLOduration=2.746030651 podStartE2EDuration="5.268256466s" podCreationTimestamp="2025-12-03 00:30:18 +0000 UTC" firstStartedPulling="2025-12-03 00:30:20.15194638 +0000 UTC m=+405.793966940" lastFinishedPulling="2025-12-03 00:30:22.674172195 +0000 UTC m=+408.316192755" observedRunningTime="2025-12-03 00:30:23.246166657 +0000 UTC m=+408.888187217" watchObservedRunningTime="2025-12-03 00:30:23.268256466 +0000 UTC m=+408.910277026" Dec 03 00:30:24 crc kubenswrapper[4912]: I1203 00:30:24.211986 4912 generic.go:334] "Generic (PLEG): container finished" podID="8e9fe039-8cbd-4a92-9b85-4a17818ade89" containerID="771573c7c8d0064fd200f0bc4af3610ac9bb5a0e571441de743613ae2cbcc784" exitCode=0 Dec 03 00:30:24 crc kubenswrapper[4912]: I1203 00:30:24.212049 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkzm2" event={"ID":"8e9fe039-8cbd-4a92-9b85-4a17818ade89","Type":"ContainerDied","Data":"771573c7c8d0064fd200f0bc4af3610ac9bb5a0e571441de743613ae2cbcc784"} Dec 03 00:30:25 crc kubenswrapper[4912]: I1203 00:30:25.220052 4912 generic.go:334] "Generic (PLEG): container finished" podID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerID="59db0e41c7e0d79491b7cca8fafd497c031b3dbf72a5d23570c46a4185ee298d" exitCode=0 Dec 03 00:30:25 crc kubenswrapper[4912]: I1203 00:30:25.220097 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xg4j" event={"ID":"770b4a36-b3be-4828-9e41-3fb24fa639f5","Type":"ContainerDied","Data":"59db0e41c7e0d79491b7cca8fafd497c031b3dbf72a5d23570c46a4185ee298d"} Dec 03 00:30:25 crc kubenswrapper[4912]: I1203 00:30:25.829063 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" podUID="a64deb49-4066-4d06-92fb-cd522ed7e496" containerName="registry" containerID="cri-o://fa7de72ef46cf4c0312c41ac3ef480da5028328b8bd677e2f50aa4931a15e6e1" gracePeriod=30 Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.230119 4912 generic.go:334] "Generic (PLEG): container finished" podID="a64deb49-4066-4d06-92fb-cd522ed7e496" containerID="fa7de72ef46cf4c0312c41ac3ef480da5028328b8bd677e2f50aa4931a15e6e1" exitCode=0 Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.230177 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" event={"ID":"a64deb49-4066-4d06-92fb-cd522ed7e496","Type":"ContainerDied","Data":"fa7de72ef46cf4c0312c41ac3ef480da5028328b8bd677e2f50aa4931a15e6e1"} Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.558090 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.705184 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a64deb49-4066-4d06-92fb-cd522ed7e496-installation-pull-secrets\") pod \"a64deb49-4066-4d06-92fb-cd522ed7e496\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.705417 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a64deb49-4066-4d06-92fb-cd522ed7e496\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.705518 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-trusted-ca\") pod \"a64deb49-4066-4d06-92fb-cd522ed7e496\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.705573 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmflw\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-kube-api-access-bmflw\") pod \"a64deb49-4066-4d06-92fb-cd522ed7e496\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.705611 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-bound-sa-token\") pod \"a64deb49-4066-4d06-92fb-cd522ed7e496\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.705666 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-tls\") pod \"a64deb49-4066-4d06-92fb-cd522ed7e496\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.705682 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a64deb49-4066-4d06-92fb-cd522ed7e496-ca-trust-extracted\") pod \"a64deb49-4066-4d06-92fb-cd522ed7e496\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.705715 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-certificates\") pod \"a64deb49-4066-4d06-92fb-cd522ed7e496\" (UID: \"a64deb49-4066-4d06-92fb-cd522ed7e496\") " Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.706397 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a64deb49-4066-4d06-92fb-cd522ed7e496" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.710416 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64deb49-4066-4d06-92fb-cd522ed7e496-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a64deb49-4066-4d06-92fb-cd522ed7e496" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.710479 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a64deb49-4066-4d06-92fb-cd522ed7e496" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.710552 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a64deb49-4066-4d06-92fb-cd522ed7e496" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.710841 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a64deb49-4066-4d06-92fb-cd522ed7e496" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.713784 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-kube-api-access-bmflw" (OuterVolumeSpecName: "kube-api-access-bmflw") pod "a64deb49-4066-4d06-92fb-cd522ed7e496" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496"). InnerVolumeSpecName "kube-api-access-bmflw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.726056 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a64deb49-4066-4d06-92fb-cd522ed7e496-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a64deb49-4066-4d06-92fb-cd522ed7e496" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.807303 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.807338 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmflw\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-kube-api-access-bmflw\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.807347 4912 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.807380 4912 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.807391 4912 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a64deb49-4066-4d06-92fb-cd522ed7e496-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.807400 4912 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a64deb49-4066-4d06-92fb-cd522ed7e496-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:26 crc kubenswrapper[4912]: I1203 00:30:26.807408 4912 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a64deb49-4066-4d06-92fb-cd522ed7e496-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 00:30:27 crc kubenswrapper[4912]: I1203 00:30:27.235924 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a64deb49-4066-4d06-92fb-cd522ed7e496" (UID: "a64deb49-4066-4d06-92fb-cd522ed7e496"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 00:30:27 crc kubenswrapper[4912]: I1203 00:30:27.240961 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" event={"ID":"a64deb49-4066-4d06-92fb-cd522ed7e496","Type":"ContainerDied","Data":"1617c385f580def901687574248c144ec5d107f4121095eebd10d7c5c5f3f184"} Dec 03 00:30:27 crc kubenswrapper[4912]: I1203 00:30:27.241020 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-vbppn" Dec 03 00:30:27 crc kubenswrapper[4912]: I1203 00:30:27.241063 4912 scope.go:117] "RemoveContainer" containerID="fa7de72ef46cf4c0312c41ac3ef480da5028328b8bd677e2f50aa4931a15e6e1" Dec 03 00:30:27 crc kubenswrapper[4912]: I1203 00:30:27.274645 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbppn"] Dec 03 00:30:27 crc kubenswrapper[4912]: I1203 00:30:27.277942 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-vbppn"] Dec 03 00:30:28 crc kubenswrapper[4912]: I1203 00:30:28.248656 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkzm2" event={"ID":"8e9fe039-8cbd-4a92-9b85-4a17818ade89","Type":"ContainerStarted","Data":"0b6b37b16dc735ef01a8c79ef12c1b3b1dd2042afe29dac3724ec054d70c1d5e"} Dec 03 00:30:28 crc kubenswrapper[4912]: I1203 00:30:28.253890 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xg4j" event={"ID":"770b4a36-b3be-4828-9e41-3fb24fa639f5","Type":"ContainerStarted","Data":"84aad5e461f43ab08600164c58f44b2840f3f4b03ce24164fddaef981f1aac14"} Dec 03 00:30:28 crc kubenswrapper[4912]: I1203 00:30:28.268953 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bkzm2" podStartSLOduration=4.388238973 podStartE2EDuration="7.26893534s" podCreationTimestamp="2025-12-03 00:30:21 +0000 UTC" firstStartedPulling="2025-12-03 00:30:22.189459781 +0000 UTC m=+407.831480341" lastFinishedPulling="2025-12-03 00:30:25.070156148 +0000 UTC m=+410.712176708" observedRunningTime="2025-12-03 00:30:28.26669641 +0000 UTC m=+413.908716990" watchObservedRunningTime="2025-12-03 00:30:28.26893534 +0000 UTC m=+413.910955900" Dec 03 00:30:28 crc kubenswrapper[4912]: I1203 00:30:28.284406 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5xg4j" podStartSLOduration=4.847802896 podStartE2EDuration="7.284385749s" podCreationTimestamp="2025-12-03 00:30:21 +0000 UTC" firstStartedPulling="2025-12-03 00:30:23.206554863 +0000 UTC m=+408.848575423" lastFinishedPulling="2025-12-03 00:30:25.643137716 +0000 UTC m=+411.285158276" observedRunningTime="2025-12-03 00:30:28.283622659 +0000 UTC m=+413.925643239" watchObservedRunningTime="2025-12-03 00:30:28.284385749 +0000 UTC m=+413.926406309" Dec 03 00:30:28 crc kubenswrapper[4912]: I1203 00:30:28.578118 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a64deb49-4066-4d06-92fb-cd522ed7e496" path="/var/lib/kubelet/pods/a64deb49-4066-4d06-92fb-cd522ed7e496/volumes" Dec 03 00:30:29 crc kubenswrapper[4912]: I1203 00:30:29.043751 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:29 crc kubenswrapper[4912]: I1203 00:30:29.044150 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:29 crc kubenswrapper[4912]: I1203 00:30:29.229841 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:29 crc kubenswrapper[4912]: I1203 00:30:29.252893 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:29 crc kubenswrapper[4912]: I1203 00:30:29.252935 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:29 crc kubenswrapper[4912]: I1203 00:30:29.290645 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:29 crc kubenswrapper[4912]: I1203 00:30:29.307134 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jfl9b" Dec 03 00:30:29 crc kubenswrapper[4912]: I1203 00:30:29.358086 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qqv59" Dec 03 00:30:31 crc kubenswrapper[4912]: I1203 00:30:31.432250 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:31 crc kubenswrapper[4912]: I1203 00:30:31.432597 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:31 crc kubenswrapper[4912]: I1203 00:30:31.483446 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:31 crc kubenswrapper[4912]: I1203 00:30:31.637039 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:31 crc kubenswrapper[4912]: I1203 00:30:31.637383 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:32 crc kubenswrapper[4912]: I1203 00:30:32.325613 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bkzm2" Dec 03 00:30:32 crc kubenswrapper[4912]: I1203 00:30:32.684553 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5xg4j" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerName="registry-server" probeResult="failure" output=< Dec 03 00:30:32 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 00:30:32 crc kubenswrapper[4912]: > Dec 03 00:30:41 crc kubenswrapper[4912]: I1203 00:30:41.680681 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:41 crc kubenswrapper[4912]: I1203 00:30:41.744031 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 00:30:48 crc kubenswrapper[4912]: I1203 00:30:48.078179 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:30:48 crc kubenswrapper[4912]: I1203 00:30:48.078794 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:30:48 crc kubenswrapper[4912]: I1203 00:30:48.078873 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:30:48 crc kubenswrapper[4912]: I1203 00:30:48.079813 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be8ecf588e7c78f1c19a77cfa713fb75d6b341f0fd910a73d8ce9d0ff688932f"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:30:48 crc kubenswrapper[4912]: I1203 00:30:48.079905 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://be8ecf588e7c78f1c19a77cfa713fb75d6b341f0fd910a73d8ce9d0ff688932f" gracePeriod=600 Dec 03 00:30:48 crc kubenswrapper[4912]: I1203 00:30:48.373287 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="be8ecf588e7c78f1c19a77cfa713fb75d6b341f0fd910a73d8ce9d0ff688932f" exitCode=0 Dec 03 00:30:48 crc kubenswrapper[4912]: I1203 00:30:48.373360 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"be8ecf588e7c78f1c19a77cfa713fb75d6b341f0fd910a73d8ce9d0ff688932f"} Dec 03 00:30:48 crc kubenswrapper[4912]: I1203 00:30:48.373396 4912 scope.go:117] "RemoveContainer" containerID="40dce610bc7b5595aac70c864a54ca1a8b29e8aa24d3884f0b195923e2ac61ed" Dec 03 00:30:49 crc kubenswrapper[4912]: I1203 00:30:49.384638 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"67d01fbca4f2a6414cad35bc8cfd7bf8c1cdf2df989831ac2d2a99ac36bfb004"} Dec 03 00:32:48 crc kubenswrapper[4912]: I1203 00:32:48.078723 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:32:48 crc kubenswrapper[4912]: I1203 00:32:48.079526 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:33:18 crc kubenswrapper[4912]: I1203 00:33:18.078343 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:33:18 crc kubenswrapper[4912]: I1203 00:33:18.078916 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:33:48 crc kubenswrapper[4912]: I1203 00:33:48.078016 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:33:48 crc kubenswrapper[4912]: I1203 00:33:48.078843 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:33:48 crc kubenswrapper[4912]: I1203 00:33:48.078931 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:33:48 crc kubenswrapper[4912]: I1203 00:33:48.079763 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67d01fbca4f2a6414cad35bc8cfd7bf8c1cdf2df989831ac2d2a99ac36bfb004"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:33:48 crc kubenswrapper[4912]: I1203 00:33:48.079833 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://67d01fbca4f2a6414cad35bc8cfd7bf8c1cdf2df989831ac2d2a99ac36bfb004" gracePeriod=600 Dec 03 00:33:48 crc kubenswrapper[4912]: I1203 00:33:48.433461 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="67d01fbca4f2a6414cad35bc8cfd7bf8c1cdf2df989831ac2d2a99ac36bfb004" exitCode=0 Dec 03 00:33:48 crc kubenswrapper[4912]: I1203 00:33:48.433734 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"67d01fbca4f2a6414cad35bc8cfd7bf8c1cdf2df989831ac2d2a99ac36bfb004"} Dec 03 00:33:48 crc kubenswrapper[4912]: I1203 00:33:48.433826 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"6db36669257e6a0bbce38d40b2b10188b2e1ca05791ecba752a07c893daa324b"} Dec 03 00:33:48 crc kubenswrapper[4912]: I1203 00:33:48.433844 4912 scope.go:117] "RemoveContainer" containerID="be8ecf588e7c78f1c19a77cfa713fb75d6b341f0fd910a73d8ce9d0ff688932f" Dec 03 00:35:48 crc kubenswrapper[4912]: I1203 00:35:48.078361 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:35:48 crc kubenswrapper[4912]: I1203 00:35:48.079545 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:36:18 crc kubenswrapper[4912]: I1203 00:36:18.078508 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:36:18 crc kubenswrapper[4912]: I1203 00:36:18.079147 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:36:18 crc kubenswrapper[4912]: I1203 00:36:18.365170 4912 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.115504 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc"] Dec 03 00:36:29 crc kubenswrapper[4912]: E1203 00:36:29.116593 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64deb49-4066-4d06-92fb-cd522ed7e496" containerName="registry" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.116610 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64deb49-4066-4d06-92fb-cd522ed7e496" containerName="registry" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.116728 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64deb49-4066-4d06-92fb-cd522ed7e496" containerName="registry" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.117672 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.120011 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.127132 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc"] Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.271501 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.271593 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.271613 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l56j4\" (UniqueName: \"kubernetes.io/projected/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-kube-api-access-l56j4\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.372840 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.372895 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.372915 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l56j4\" (UniqueName: \"kubernetes.io/projected/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-kube-api-access-l56j4\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.373420 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.373548 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.400236 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l56j4\" (UniqueName: \"kubernetes.io/projected/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-kube-api-access-l56j4\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.436053 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:29 crc kubenswrapper[4912]: I1203 00:36:29.653577 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc"] Dec 03 00:36:30 crc kubenswrapper[4912]: I1203 00:36:30.510044 4912 generic.go:334] "Generic (PLEG): container finished" podID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerID="e8fffb723750d813ac252706cc1e4559647f251d72e979e66dc0807b8edc2ddd" exitCode=0 Dec 03 00:36:30 crc kubenswrapper[4912]: I1203 00:36:30.510107 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" event={"ID":"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3","Type":"ContainerDied","Data":"e8fffb723750d813ac252706cc1e4559647f251d72e979e66dc0807b8edc2ddd"} Dec 03 00:36:30 crc kubenswrapper[4912]: I1203 00:36:30.510589 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" event={"ID":"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3","Type":"ContainerStarted","Data":"9be0d5df0756083ac2f21d2b9bf05841d0644070533e723bc96fc90b23f8c7a1"} Dec 03 00:36:30 crc kubenswrapper[4912]: I1203 00:36:30.511936 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.465945 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d9mk6"] Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.467270 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.483275 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d9mk6"] Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.614303 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-catalog-content\") pod \"redhat-operators-d9mk6\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.614381 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-utilities\") pod \"redhat-operators-d9mk6\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.614461 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtgz9\" (UniqueName: \"kubernetes.io/projected/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-kube-api-access-vtgz9\") pod \"redhat-operators-d9mk6\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.716118 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtgz9\" (UniqueName: \"kubernetes.io/projected/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-kube-api-access-vtgz9\") pod \"redhat-operators-d9mk6\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.716201 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-catalog-content\") pod \"redhat-operators-d9mk6\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.716271 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-utilities\") pod \"redhat-operators-d9mk6\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.717556 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-catalog-content\") pod \"redhat-operators-d9mk6\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.718661 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-utilities\") pod \"redhat-operators-d9mk6\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.736506 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtgz9\" (UniqueName: \"kubernetes.io/projected/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-kube-api-access-vtgz9\") pod \"redhat-operators-d9mk6\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:31 crc kubenswrapper[4912]: I1203 00:36:31.784793 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:32 crc kubenswrapper[4912]: I1203 00:36:32.001207 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d9mk6"] Dec 03 00:36:32 crc kubenswrapper[4912]: W1203 00:36:32.007318 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc068c38b_d1c8_4223_b7b6_4da7f113f2ea.slice/crio-e0423543ca789c6c494698f40c9d1b48499968c5442ed6100887b296aaa65e77 WatchSource:0}: Error finding container e0423543ca789c6c494698f40c9d1b48499968c5442ed6100887b296aaa65e77: Status 404 returned error can't find the container with id e0423543ca789c6c494698f40c9d1b48499968c5442ed6100887b296aaa65e77 Dec 03 00:36:32 crc kubenswrapper[4912]: I1203 00:36:32.525026 4912 generic.go:334] "Generic (PLEG): container finished" podID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerID="1f112caed7d4e77dfbf1ddb135b76998972b3e9b28e6a2b8166ec3686587cb68" exitCode=0 Dec 03 00:36:32 crc kubenswrapper[4912]: I1203 00:36:32.525103 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" event={"ID":"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3","Type":"ContainerDied","Data":"1f112caed7d4e77dfbf1ddb135b76998972b3e9b28e6a2b8166ec3686587cb68"} Dec 03 00:36:32 crc kubenswrapper[4912]: I1203 00:36:32.527204 4912 generic.go:334] "Generic (PLEG): container finished" podID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerID="05e153b239407ed28bfa1130bfc504740a0782c86e25016e30a20c322cd18456" exitCode=0 Dec 03 00:36:32 crc kubenswrapper[4912]: I1203 00:36:32.527242 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9mk6" event={"ID":"c068c38b-d1c8-4223-b7b6-4da7f113f2ea","Type":"ContainerDied","Data":"05e153b239407ed28bfa1130bfc504740a0782c86e25016e30a20c322cd18456"} Dec 03 00:36:32 crc kubenswrapper[4912]: I1203 00:36:32.527266 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9mk6" event={"ID":"c068c38b-d1c8-4223-b7b6-4da7f113f2ea","Type":"ContainerStarted","Data":"e0423543ca789c6c494698f40c9d1b48499968c5442ed6100887b296aaa65e77"} Dec 03 00:36:33 crc kubenswrapper[4912]: I1203 00:36:33.535966 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9mk6" event={"ID":"c068c38b-d1c8-4223-b7b6-4da7f113f2ea","Type":"ContainerStarted","Data":"114696955b566cf1731a42db40c00a18e8188b6fbfa076e1104a0586f5681cda"} Dec 03 00:36:33 crc kubenswrapper[4912]: I1203 00:36:33.539049 4912 generic.go:334] "Generic (PLEG): container finished" podID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerID="e86c14ad1c358b38689ed6a7666e792b1affb435c854fd20f54ee804736a21e2" exitCode=0 Dec 03 00:36:33 crc kubenswrapper[4912]: I1203 00:36:33.539091 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" event={"ID":"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3","Type":"ContainerDied","Data":"e86c14ad1c358b38689ed6a7666e792b1affb435c854fd20f54ee804736a21e2"} Dec 03 00:36:34 crc kubenswrapper[4912]: I1203 00:36:34.547620 4912 generic.go:334] "Generic (PLEG): container finished" podID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerID="114696955b566cf1731a42db40c00a18e8188b6fbfa076e1104a0586f5681cda" exitCode=0 Dec 03 00:36:34 crc kubenswrapper[4912]: I1203 00:36:34.547734 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9mk6" event={"ID":"c068c38b-d1c8-4223-b7b6-4da7f113f2ea","Type":"ContainerDied","Data":"114696955b566cf1731a42db40c00a18e8188b6fbfa076e1104a0586f5681cda"} Dec 03 00:36:34 crc kubenswrapper[4912]: I1203 00:36:34.882400 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:34 crc kubenswrapper[4912]: I1203 00:36:34.956652 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-util\") pod \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " Dec 03 00:36:34 crc kubenswrapper[4912]: I1203 00:36:34.956756 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-bundle\") pod \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " Dec 03 00:36:34 crc kubenswrapper[4912]: I1203 00:36:34.956801 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l56j4\" (UniqueName: \"kubernetes.io/projected/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-kube-api-access-l56j4\") pod \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\" (UID: \"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3\") " Dec 03 00:36:34 crc kubenswrapper[4912]: I1203 00:36:34.958977 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-bundle" (OuterVolumeSpecName: "bundle") pod "e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" (UID: "e3e0e84c-fd2b-41ac-adda-0c32d431e5f3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:36:34 crc kubenswrapper[4912]: I1203 00:36:34.961724 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-kube-api-access-l56j4" (OuterVolumeSpecName: "kube-api-access-l56j4") pod "e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" (UID: "e3e0e84c-fd2b-41ac-adda-0c32d431e5f3"). InnerVolumeSpecName "kube-api-access-l56j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:36:34 crc kubenswrapper[4912]: I1203 00:36:34.970266 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-util" (OuterVolumeSpecName: "util") pod "e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" (UID: "e3e0e84c-fd2b-41ac-adda-0c32d431e5f3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:36:35 crc kubenswrapper[4912]: I1203 00:36:35.057953 4912 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:35 crc kubenswrapper[4912]: I1203 00:36:35.057996 4912 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:35 crc kubenswrapper[4912]: I1203 00:36:35.058010 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l56j4\" (UniqueName: \"kubernetes.io/projected/e3e0e84c-fd2b-41ac-adda-0c32d431e5f3-kube-api-access-l56j4\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:35 crc kubenswrapper[4912]: I1203 00:36:35.561364 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" event={"ID":"e3e0e84c-fd2b-41ac-adda-0c32d431e5f3","Type":"ContainerDied","Data":"9be0d5df0756083ac2f21d2b9bf05841d0644070533e723bc96fc90b23f8c7a1"} Dec 03 00:36:35 crc kubenswrapper[4912]: I1203 00:36:35.561748 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9be0d5df0756083ac2f21d2b9bf05841d0644070533e723bc96fc90b23f8c7a1" Dec 03 00:36:35 crc kubenswrapper[4912]: I1203 00:36:35.561399 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc" Dec 03 00:36:35 crc kubenswrapper[4912]: I1203 00:36:35.564024 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9mk6" event={"ID":"c068c38b-d1c8-4223-b7b6-4da7f113f2ea","Type":"ContainerStarted","Data":"5b901826c74c1f1d6802d6aa2add0af50a30cd07efb6bec7b8f85e41bc3e00d9"} Dec 03 00:36:35 crc kubenswrapper[4912]: I1203 00:36:35.581360 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d9mk6" podStartSLOduration=2.040555172 podStartE2EDuration="4.58133043s" podCreationTimestamp="2025-12-03 00:36:31 +0000 UTC" firstStartedPulling="2025-12-03 00:36:32.528267706 +0000 UTC m=+778.170288276" lastFinishedPulling="2025-12-03 00:36:35.069042974 +0000 UTC m=+780.711063534" observedRunningTime="2025-12-03 00:36:35.580510217 +0000 UTC m=+781.222530807" watchObservedRunningTime="2025-12-03 00:36:35.58133043 +0000 UTC m=+781.223351000" Dec 03 00:36:40 crc kubenswrapper[4912]: I1203 00:36:40.174417 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f29xt"] Dec 03 00:36:40 crc kubenswrapper[4912]: I1203 00:36:40.175096 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovn-controller" containerID="cri-o://81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24" gracePeriod=30 Dec 03 00:36:40 crc kubenswrapper[4912]: I1203 00:36:40.175154 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="nbdb" containerID="cri-o://730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1" gracePeriod=30 Dec 03 00:36:40 crc kubenswrapper[4912]: I1203 00:36:40.175205 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kube-rbac-proxy-node" containerID="cri-o://649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45" gracePeriod=30 Dec 03 00:36:40 crc kubenswrapper[4912]: I1203 00:36:40.175191 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7" gracePeriod=30 Dec 03 00:36:40 crc kubenswrapper[4912]: I1203 00:36:40.175243 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovn-acl-logging" containerID="cri-o://d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67" gracePeriod=30 Dec 03 00:36:40 crc kubenswrapper[4912]: I1203 00:36:40.175275 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="northd" containerID="cri-o://ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa" gracePeriod=30 Dec 03 00:36:40 crc kubenswrapper[4912]: I1203 00:36:40.175479 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="sbdb" containerID="cri-o://d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1" gracePeriod=30 Dec 03 00:36:40 crc kubenswrapper[4912]: I1203 00:36:40.211380 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" containerID="cri-o://591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3" gracePeriod=30 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.494297 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/3.log" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.496526 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovn-acl-logging/0.log" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.497072 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovn-controller/0.log" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.497579 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.587890 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hcxn8"] Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588124 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kubecfg-setup" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588139 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kubecfg-setup" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588148 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerName="pull" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588157 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerName="pull" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588166 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588173 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588183 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerName="extract" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588189 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerName="extract" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588196 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="nbdb" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588202 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="nbdb" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588213 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588220 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588230 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="northd" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588236 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="northd" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588243 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588249 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588257 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588263 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588270 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588275 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588283 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="sbdb" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588289 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="sbdb" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588297 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovn-acl-logging" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588303 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovn-acl-logging" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588311 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerName="util" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588316 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerName="util" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588325 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovn-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588330 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovn-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588338 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588344 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 00:36:41 crc kubenswrapper[4912]: E1203 00:36:41.588352 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kube-rbac-proxy-node" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588358 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kube-rbac-proxy-node" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588457 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588466 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588472 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="sbdb" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588480 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588487 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovn-acl-logging" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588497 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e0e84c-fd2b-41ac-adda-0c32d431e5f3" containerName="extract" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588504 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="kube-rbac-proxy-node" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588511 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="northd" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588517 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="nbdb" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588523 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588530 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovn-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588704 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.588716 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerName="ovnkube-controller" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.590224 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.603039 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2wz8_8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77/kube-multus/2.log" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.603735 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2wz8_8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77/kube-multus/1.log" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.603780 4912 generic.go:334] "Generic (PLEG): container finished" podID="8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77" containerID="892e7bbdead6a107c57ae944c2e00a9f798c609b0c8b755fb63c17c64a088a49" exitCode=2 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.603859 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2wz8" event={"ID":"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77","Type":"ContainerDied","Data":"892e7bbdead6a107c57ae944c2e00a9f798c609b0c8b755fb63c17c64a088a49"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.603905 4912 scope.go:117] "RemoveContainer" containerID="62fa436b5f198bad55c734f6dd68f3cea28dad6f7bb300998e3cf4be5f24270e" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.604571 4912 scope.go:117] "RemoveContainer" containerID="892e7bbdead6a107c57ae944c2e00a9f798c609b0c8b755fb63c17c64a088a49" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.609804 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovnkube-controller/3.log" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.612789 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovn-acl-logging/0.log" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.615073 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f29xt_7d5e7779-476a-4b6f-8153-7fd40fab8d90/ovn-controller/0.log" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.615674 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3" exitCode=0 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.615756 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1" exitCode=0 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.615829 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1" exitCode=0 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.615897 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa" exitCode=0 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.615956 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7" exitCode=0 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616027 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45" exitCode=0 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616108 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67" exitCode=143 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616174 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" containerID="81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24" exitCode=143 Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616254 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616349 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616459 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616539 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616628 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616704 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616774 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616846 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616905 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.616971 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617031 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617089 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617146 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617205 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617265 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617328 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617394 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617480 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617550 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617607 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617681 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617743 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617839 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617913 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617967 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618013 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618057 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618109 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618179 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618249 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618296 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618341 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618488 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618545 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618595 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618639 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618687 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618736 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618784 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" event={"ID":"7d5e7779-476a-4b6f-8153-7fd40fab8d90","Type":"ContainerDied","Data":"35a3a7e8002b02de59d575e09d842d32a17f0e9d464e195bcd8bff14e04d9d8f"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618837 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618884 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618928 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.618976 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.619025 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.619079 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.619147 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.619195 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.619242 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.619299 4912 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01"} Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.617632 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f29xt" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.640339 4912 scope.go:117] "RemoveContainer" containerID="591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.650821 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-netd\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651092 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-openvswitch\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651188 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovn-node-metrics-cert\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651266 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs5bf\" (UniqueName: \"kubernetes.io/projected/7d5e7779-476a-4b6f-8153-7fd40fab8d90-kube-api-access-hs5bf\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651330 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-var-lib-openvswitch\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651409 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-env-overrides\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651500 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-ovn-kubernetes\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651565 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-script-lib\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651626 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-kubelet\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651739 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-systemd-units\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651802 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-config\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651862 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-bin\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.651927 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-log-socket\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652020 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-etc-openvswitch\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652118 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-var-lib-cni-networks-ovn-kubernetes\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652187 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-ovn\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652260 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-systemd\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652326 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-netns\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652416 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-slash\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652515 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-node-log\") pod \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\" (UID: \"7d5e7779-476a-4b6f-8153-7fd40fab8d90\") " Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652712 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-env-overrides\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652798 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-run-ovn-kubernetes\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652868 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-run-ovn\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.652935 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-run-openvswitch\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653030 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653110 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-var-lib-openvswitch\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653175 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-ovnkube-config\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653242 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-etc-openvswitch\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653307 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-slash\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653374 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-cni-netd\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653453 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-log-socket\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653538 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-ovn-node-metrics-cert\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653620 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-kubelet\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653704 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scsgl\" (UniqueName: \"kubernetes.io/projected/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-kube-api-access-scsgl\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653787 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-ovnkube-script-lib\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653869 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-run-systemd\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653931 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-run-netns\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.653988 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-node-log\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.654066 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-systemd-units\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.654134 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-cni-bin\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.654263 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.654334 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.655737 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.655836 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.656260 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.656368 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.656772 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.656869 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.656937 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.657303 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.657390 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.657500 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-log-socket" (OuterVolumeSpecName: "log-socket") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.657568 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-slash" (OuterVolumeSpecName: "host-slash") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.657640 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.657718 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.662502 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.662553 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-node-log" (OuterVolumeSpecName: "node-log") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.663497 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.679998 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5e7779-476a-4b6f-8153-7fd40fab8d90-kube-api-access-hs5bf" (OuterVolumeSpecName: "kube-api-access-hs5bf") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "kube-api-access-hs5bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.690055 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "7d5e7779-476a-4b6f-8153-7fd40fab8d90" (UID: "7d5e7779-476a-4b6f-8153-7fd40fab8d90"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.705622 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.758058 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-ovn-node-metrics-cert\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.758315 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-kubelet\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.758420 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scsgl\" (UniqueName: \"kubernetes.io/projected/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-kube-api-access-scsgl\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.758583 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-ovnkube-script-lib\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.758691 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-run-systemd\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.758776 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-run-netns\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.758841 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-node-log\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.758913 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-systemd-units\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.758990 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-cni-bin\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.759080 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-env-overrides\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.759325 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-run-ovn-kubernetes\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.759408 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-run-ovn\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.759500 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-run-openvswitch\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.759644 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.759772 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-var-lib-openvswitch\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.763663 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-ovnkube-config\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.764200 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-etc-openvswitch\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.764280 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-slash\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.764355 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-cni-netd\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.764485 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-log-socket\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.764874 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-env-overrides\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.760974 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-cni-bin\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.764190 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-var-lib-openvswitch\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.765343 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-log-socket\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.765352 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-run-ovn-kubernetes\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.765580 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-run-openvswitch\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.765715 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-kubelet\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.767572 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-run-ovn\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.767606 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-etc-openvswitch\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.767800 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768456 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-slash\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768535 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-cni-netd\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768641 4912 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768659 4912 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768677 4912 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768710 4912 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768722 4912 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768732 4912 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768745 4912 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768754 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs5bf\" (UniqueName: \"kubernetes.io/projected/7d5e7779-476a-4b6f-8153-7fd40fab8d90-kube-api-access-hs5bf\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768763 4912 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768772 4912 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768784 4912 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768793 4912 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768803 4912 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768812 4912 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768823 4912 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7d5e7779-476a-4b6f-8153-7fd40fab8d90-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768856 4912 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768865 4912 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768874 4912 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768886 4912 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768895 4912 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7d5e7779-476a-4b6f-8153-7fd40fab8d90-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.768911 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-run-systemd\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.769105 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-ovnkube-config\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.769149 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-node-log\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.769176 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-host-run-netns\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.769201 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-systemd-units\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.786402 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.788672 4912 scope.go:117] "RemoveContainer" containerID="d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.789278 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.789513 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-ovnkube-script-lib\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.814299 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-ovn-node-metrics-cert\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.815454 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scsgl\" (UniqueName: \"kubernetes.io/projected/2b7ed935-b2f4-470a-ad77-4a2db9011ca6-kube-api-access-scsgl\") pod \"ovnkube-node-hcxn8\" (UID: \"2b7ed935-b2f4-470a-ad77-4a2db9011ca6\") " pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.852712 4912 scope.go:117] "RemoveContainer" containerID="730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.873377 4912 scope.go:117] "RemoveContainer" containerID="ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.888390 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.895831 4912 scope.go:117] "RemoveContainer" containerID="604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.906872 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.936704 4912 scope.go:117] "RemoveContainer" containerID="649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.962000 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f29xt"] Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.968666 4912 scope.go:117] "RemoveContainer" containerID="d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67" Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.970246 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f29xt"] Dec 03 00:36:41 crc kubenswrapper[4912]: I1203 00:36:41.994299 4912 scope.go:117] "RemoveContainer" containerID="81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.011599 4912 scope.go:117] "RemoveContainer" containerID="71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.047703 4912 scope.go:117] "RemoveContainer" containerID="591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.048276 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3\": container with ID starting with 591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3 not found: ID does not exist" containerID="591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.048324 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} err="failed to get container status \"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3\": rpc error: code = NotFound desc = could not find container \"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3\": container with ID starting with 591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.048355 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.048741 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\": container with ID starting with 9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff not found: ID does not exist" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.048772 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} err="failed to get container status \"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\": rpc error: code = NotFound desc = could not find container \"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\": container with ID starting with 9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.048798 4912 scope.go:117] "RemoveContainer" containerID="d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.049207 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\": container with ID starting with d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1 not found: ID does not exist" containerID="d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.049228 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} err="failed to get container status \"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\": rpc error: code = NotFound desc = could not find container \"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\": container with ID starting with d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.049245 4912 scope.go:117] "RemoveContainer" containerID="730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.049599 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\": container with ID starting with 730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1 not found: ID does not exist" containerID="730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.049623 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} err="failed to get container status \"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\": rpc error: code = NotFound desc = could not find container \"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\": container with ID starting with 730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.049642 4912 scope.go:117] "RemoveContainer" containerID="ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.050249 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\": container with ID starting with ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa not found: ID does not exist" containerID="ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.050299 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} err="failed to get container status \"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\": rpc error: code = NotFound desc = could not find container \"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\": container with ID starting with ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.050316 4912 scope.go:117] "RemoveContainer" containerID="604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.052668 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\": container with ID starting with 604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7 not found: ID does not exist" containerID="604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.052743 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} err="failed to get container status \"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\": rpc error: code = NotFound desc = could not find container \"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\": container with ID starting with 604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.052794 4912 scope.go:117] "RemoveContainer" containerID="649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.053288 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\": container with ID starting with 649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45 not found: ID does not exist" containerID="649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.053331 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} err="failed to get container status \"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\": rpc error: code = NotFound desc = could not find container \"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\": container with ID starting with 649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.053357 4912 scope.go:117] "RemoveContainer" containerID="d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.053827 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\": container with ID starting with d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67 not found: ID does not exist" containerID="d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.053856 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} err="failed to get container status \"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\": rpc error: code = NotFound desc = could not find container \"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\": container with ID starting with d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.053873 4912 scope.go:117] "RemoveContainer" containerID="81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.054074 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\": container with ID starting with 81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24 not found: ID does not exist" containerID="81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.054096 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} err="failed to get container status \"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\": rpc error: code = NotFound desc = could not find container \"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\": container with ID starting with 81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.054115 4912 scope.go:117] "RemoveContainer" containerID="71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01" Dec 03 00:36:42 crc kubenswrapper[4912]: E1203 00:36:42.054328 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\": container with ID starting with 71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01 not found: ID does not exist" containerID="71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.054348 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01"} err="failed to get container status \"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\": rpc error: code = NotFound desc = could not find container \"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\": container with ID starting with 71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.054362 4912 scope.go:117] "RemoveContainer" containerID="591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.054598 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} err="failed to get container status \"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3\": rpc error: code = NotFound desc = could not find container \"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3\": container with ID starting with 591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.054618 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.054838 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} err="failed to get container status \"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\": rpc error: code = NotFound desc = could not find container \"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\": container with ID starting with 9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.054862 4912 scope.go:117] "RemoveContainer" containerID="d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.055093 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} err="failed to get container status \"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\": rpc error: code = NotFound desc = could not find container \"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\": container with ID starting with d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.055116 4912 scope.go:117] "RemoveContainer" containerID="730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.055366 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} err="failed to get container status \"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\": rpc error: code = NotFound desc = could not find container \"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\": container with ID starting with 730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.055389 4912 scope.go:117] "RemoveContainer" containerID="ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.055688 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} err="failed to get container status \"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\": rpc error: code = NotFound desc = could not find container \"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\": container with ID starting with ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.055713 4912 scope.go:117] "RemoveContainer" containerID="604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.055983 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} err="failed to get container status \"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\": rpc error: code = NotFound desc = could not find container \"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\": container with ID starting with 604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.056052 4912 scope.go:117] "RemoveContainer" containerID="649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.056246 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} err="failed to get container status \"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\": rpc error: code = NotFound desc = could not find container \"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\": container with ID starting with 649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.056265 4912 scope.go:117] "RemoveContainer" containerID="d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.056479 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} err="failed to get container status \"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\": rpc error: code = NotFound desc = could not find container \"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\": container with ID starting with d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.056503 4912 scope.go:117] "RemoveContainer" containerID="81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.056714 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} err="failed to get container status \"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\": rpc error: code = NotFound desc = could not find container \"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\": container with ID starting with 81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.056740 4912 scope.go:117] "RemoveContainer" containerID="71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.056941 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01"} err="failed to get container status \"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\": rpc error: code = NotFound desc = could not find container \"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\": container with ID starting with 71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.056964 4912 scope.go:117] "RemoveContainer" containerID="591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.058816 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} err="failed to get container status \"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3\": rpc error: code = NotFound desc = could not find container \"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3\": container with ID starting with 591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.058866 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.059182 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} err="failed to get container status \"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\": rpc error: code = NotFound desc = could not find container \"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\": container with ID starting with 9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.059208 4912 scope.go:117] "RemoveContainer" containerID="d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.059380 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} err="failed to get container status \"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\": rpc error: code = NotFound desc = could not find container \"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\": container with ID starting with d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.059401 4912 scope.go:117] "RemoveContainer" containerID="730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.059611 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} err="failed to get container status \"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\": rpc error: code = NotFound desc = could not find container \"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\": container with ID starting with 730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.059636 4912 scope.go:117] "RemoveContainer" containerID="ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.059814 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} err="failed to get container status \"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\": rpc error: code = NotFound desc = could not find container \"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\": container with ID starting with ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.059835 4912 scope.go:117] "RemoveContainer" containerID="604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.060031 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} err="failed to get container status \"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\": rpc error: code = NotFound desc = could not find container \"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\": container with ID starting with 604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.060052 4912 scope.go:117] "RemoveContainer" containerID="649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.060249 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} err="failed to get container status \"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\": rpc error: code = NotFound desc = could not find container \"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\": container with ID starting with 649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.060275 4912 scope.go:117] "RemoveContainer" containerID="d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.060484 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} err="failed to get container status \"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\": rpc error: code = NotFound desc = could not find container \"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\": container with ID starting with d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.060502 4912 scope.go:117] "RemoveContainer" containerID="81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.060771 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} err="failed to get container status \"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\": rpc error: code = NotFound desc = could not find container \"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\": container with ID starting with 81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.060797 4912 scope.go:117] "RemoveContainer" containerID="71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061016 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01"} err="failed to get container status \"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\": rpc error: code = NotFound desc = could not find container \"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\": container with ID starting with 71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061037 4912 scope.go:117] "RemoveContainer" containerID="591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061214 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3"} err="failed to get container status \"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3\": rpc error: code = NotFound desc = could not find container \"591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3\": container with ID starting with 591ffe3a8582a7104bf94c96a53835d4682055658e91c5150830116fbf2c6ee3 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061228 4912 scope.go:117] "RemoveContainer" containerID="9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061382 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff"} err="failed to get container status \"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\": rpc error: code = NotFound desc = could not find container \"9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff\": container with ID starting with 9027b0b814dc0a31000434e989141d564b6f51b53d4db9c96f9bf68d56c529ff not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061395 4912 scope.go:117] "RemoveContainer" containerID="d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061592 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1"} err="failed to get container status \"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\": rpc error: code = NotFound desc = could not find container \"d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1\": container with ID starting with d7021c845025207c19dc9d74e68c8c756cc7168b0606dad7cad24ac240348fc1 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061609 4912 scope.go:117] "RemoveContainer" containerID="730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061783 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1"} err="failed to get container status \"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\": rpc error: code = NotFound desc = could not find container \"730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1\": container with ID starting with 730c7b70ccac4fdfc4a3c5648f11444367785da4f2c134afed0eb528b0414cc1 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061797 4912 scope.go:117] "RemoveContainer" containerID="ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061947 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa"} err="failed to get container status \"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\": rpc error: code = NotFound desc = could not find container \"ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa\": container with ID starting with ff047896b703fd7e40dd76f509c01583bfbf689fa4858904b2c24af3b896b9fa not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.061960 4912 scope.go:117] "RemoveContainer" containerID="604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.062107 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7"} err="failed to get container status \"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\": rpc error: code = NotFound desc = could not find container \"604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7\": container with ID starting with 604a1344c670b0301e44d41c8e72a13f3a1ae1fdc70799072b64a33ab9ee21b7 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.062123 4912 scope.go:117] "RemoveContainer" containerID="649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.062314 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45"} err="failed to get container status \"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\": rpc error: code = NotFound desc = could not find container \"649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45\": container with ID starting with 649c1e9655da2a6e22457be18abda3b8e5124d48fe88e51a51132a50f0c16b45 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.062336 4912 scope.go:117] "RemoveContainer" containerID="d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.062559 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67"} err="failed to get container status \"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\": rpc error: code = NotFound desc = could not find container \"d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67\": container with ID starting with d89e4a7e003d28ebf56998d4cdada935b99b7a5f5f92cac5cf191aecdebfbc67 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.062592 4912 scope.go:117] "RemoveContainer" containerID="81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.064339 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24"} err="failed to get container status \"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\": rpc error: code = NotFound desc = could not find container \"81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24\": container with ID starting with 81b1ae6b614558b845789caa5908d4fe7aff3d0ddf5634a56984f1115f6b3a24 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.064360 4912 scope.go:117] "RemoveContainer" containerID="71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.064742 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01"} err="failed to get container status \"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\": rpc error: code = NotFound desc = could not find container \"71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01\": container with ID starting with 71b66c0de65891464698fbdcb232ffbcd6d16693d4d4719ec3f6f3fd751f6a01 not found: ID does not exist" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.579279 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d5e7779-476a-4b6f-8153-7fd40fab8d90" path="/var/lib/kubelet/pods/7d5e7779-476a-4b6f-8153-7fd40fab8d90/volumes" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.622976 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m2wz8_8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77/kube-multus/2.log" Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.623062 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m2wz8" event={"ID":"8260bfbd-d3da-4ea9-b9ad-1ac726fe0a77","Type":"ContainerStarted","Data":"da3a7cfeda481d6f488a0f9b3da9f2ff38c2cd66a8b17690cd919c11b1a3b6e0"} Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.625168 4912 generic.go:334] "Generic (PLEG): container finished" podID="2b7ed935-b2f4-470a-ad77-4a2db9011ca6" containerID="7978a627a9edbc463b676a920067d3b7919916b40552128f37beb59c47350536" exitCode=0 Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.625240 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerDied","Data":"7978a627a9edbc463b676a920067d3b7919916b40552128f37beb59c47350536"} Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.625266 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerStarted","Data":"a8a91f07aeae377b641bde5d5a566152b0a66c22686a37fe76fc714473723674"} Dec 03 00:36:42 crc kubenswrapper[4912]: I1203 00:36:42.703439 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:43 crc kubenswrapper[4912]: I1203 00:36:43.636015 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerStarted","Data":"385599aff4663b9516727dcb285c07cbe4e4feb17686325b7dd76d369cd24fd3"} Dec 03 00:36:43 crc kubenswrapper[4912]: I1203 00:36:43.636585 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerStarted","Data":"42514fcab9602e5c55fbb31e2c19d7a23f2aa98ae8cba9f89a625bdda53953ba"} Dec 03 00:36:43 crc kubenswrapper[4912]: I1203 00:36:43.636600 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerStarted","Data":"d518ac85291297bd272fc5b6a58bc1f5f64de964a367a0e4bb1e9a2c73258341"} Dec 03 00:36:43 crc kubenswrapper[4912]: I1203 00:36:43.636610 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerStarted","Data":"a6e1331f90748f8c641adbeae2a0b9c46c1ccca13c5dac0a4521c04b4a103a2c"} Dec 03 00:36:43 crc kubenswrapper[4912]: I1203 00:36:43.636619 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerStarted","Data":"ca31aecdaa2f56ef6acd2049ec2a0ae96daef5b5dfa45742f7015e88c0688238"} Dec 03 00:36:43 crc kubenswrapper[4912]: I1203 00:36:43.636630 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerStarted","Data":"6a3038a5766f80d79072a84e909530aca3efd092fad24127ee2bc3a506e24279"} Dec 03 00:36:45 crc kubenswrapper[4912]: I1203 00:36:45.193151 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d9mk6"] Dec 03 00:36:45 crc kubenswrapper[4912]: I1203 00:36:45.653872 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d9mk6" podUID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerName="registry-server" containerID="cri-o://5b901826c74c1f1d6802d6aa2add0af50a30cd07efb6bec7b8f85e41bc3e00d9" gracePeriod=2 Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.135885 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8"] Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.136597 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.138203 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-s827w" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.138264 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.140368 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.264619 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-478wd\" (UniqueName: \"kubernetes.io/projected/377b37d8-070e-4463-9ea8-c698c4e289a4-kube-api-access-478wd\") pod \"obo-prometheus-operator-668cf9dfbb-spkm8\" (UID: \"377b37d8-070e-4463-9ea8-c698c4e289a4\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.364162 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2"] Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.369173 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.367514 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-478wd\" (UniqueName: \"kubernetes.io/projected/377b37d8-070e-4463-9ea8-c698c4e289a4-kube-api-access-478wd\") pod \"obo-prometheus-operator-668cf9dfbb-spkm8\" (UID: \"377b37d8-070e-4463-9ea8-c698c4e289a4\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.377958 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.378148 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-2w2gt" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.385740 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5"] Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.387374 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.450749 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-478wd\" (UniqueName: \"kubernetes.io/projected/377b37d8-070e-4463-9ea8-c698c4e289a4-kube-api-access-478wd\") pod \"obo-prometheus-operator-668cf9dfbb-spkm8\" (UID: \"377b37d8-070e-4463-9ea8-c698c4e289a4\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.465867 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.471268 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2\" (UID: \"45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.471340 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/194f5401-404a-43b8-b826-3df930390630-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5\" (UID: \"194f5401-404a-43b8-b826-3df930390630\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.471378 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2\" (UID: \"45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.471414 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/194f5401-404a-43b8-b826-3df930390630-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5\" (UID: \"194f5401-404a-43b8-b826-3df930390630\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.524650 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators_377b37d8-070e-4463-9ea8-c698c4e289a4_0(95033550e8e34e383becc2869f7c2c0a1500ddf3b9237f6d39101f01273ac1f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.524755 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators_377b37d8-070e-4463-9ea8-c698c4e289a4_0(95033550e8e34e383becc2869f7c2c0a1500ddf3b9237f6d39101f01273ac1f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.524781 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators_377b37d8-070e-4463-9ea8-c698c4e289a4_0(95033550e8e34e383becc2869f7c2c0a1500ddf3b9237f6d39101f01273ac1f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.524838 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators(377b37d8-070e-4463-9ea8-c698c4e289a4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators(377b37d8-070e-4463-9ea8-c698c4e289a4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators_377b37d8-070e-4463-9ea8-c698c4e289a4_0(95033550e8e34e383becc2869f7c2c0a1500ddf3b9237f6d39101f01273ac1f5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" podUID="377b37d8-070e-4463-9ea8-c698c4e289a4" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.530191 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-bsscx"] Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.531158 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.541888 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.542878 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-qgqww" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.572239 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/194f5401-404a-43b8-b826-3df930390630-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5\" (UID: \"194f5401-404a-43b8-b826-3df930390630\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.572321 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2\" (UID: \"45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.572400 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/194f5401-404a-43b8-b826-3df930390630-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5\" (UID: \"194f5401-404a-43b8-b826-3df930390630\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.572468 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2\" (UID: \"45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.577736 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2\" (UID: \"45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.582265 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/194f5401-404a-43b8-b826-3df930390630-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5\" (UID: \"194f5401-404a-43b8-b826-3df930390630\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.585043 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2\" (UID: \"45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.593322 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/194f5401-404a-43b8-b826-3df930390630-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5\" (UID: \"194f5401-404a-43b8-b826-3df930390630\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.666285 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerStarted","Data":"4cb401416e5e47c1741d8544d909912d61c4628c21c6811f129a1f352869b2ab"} Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.668486 4912 generic.go:334] "Generic (PLEG): container finished" podID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerID="5b901826c74c1f1d6802d6aa2add0af50a30cd07efb6bec7b8f85e41bc3e00d9" exitCode=0 Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.668530 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9mk6" event={"ID":"c068c38b-d1c8-4223-b7b6-4da7f113f2ea","Type":"ContainerDied","Data":"5b901826c74c1f1d6802d6aa2add0af50a30cd07efb6bec7b8f85e41bc3e00d9"} Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.673561 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4a067c7-ef55-4185-8315-a61b5bfc8e22-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-bsscx\" (UID: \"f4a067c7-ef55-4185-8315-a61b5bfc8e22\") " pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.673646 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmmx5\" (UniqueName: \"kubernetes.io/projected/f4a067c7-ef55-4185-8315-a61b5bfc8e22-kube-api-access-gmmx5\") pod \"observability-operator-d8bb48f5d-bsscx\" (UID: \"f4a067c7-ef55-4185-8315-a61b5bfc8e22\") " pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.678549 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-trfdr"] Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.679587 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.683236 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-4cg6f" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.721373 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.738635 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0_0(feec5b1305dc67f565320bb8fdb447d4234c12080fe42188eec5f81dba1a4320): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.738704 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0_0(feec5b1305dc67f565320bb8fdb447d4234c12080fe42188eec5f81dba1a4320): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.738726 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0_0(feec5b1305dc67f565320bb8fdb447d4234c12080fe42188eec5f81dba1a4320): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.738775 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators(45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators(45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0_0(feec5b1305dc67f565320bb8fdb447d4234c12080fe42188eec5f81dba1a4320): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" podUID="45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.769061 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.774952 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4a067c7-ef55-4185-8315-a61b5bfc8e22-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-bsscx\" (UID: \"f4a067c7-ef55-4185-8315-a61b5bfc8e22\") " pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.774990 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmmx5\" (UniqueName: \"kubernetes.io/projected/f4a067c7-ef55-4185-8315-a61b5bfc8e22-kube-api-access-gmmx5\") pod \"observability-operator-d8bb48f5d-bsscx\" (UID: \"f4a067c7-ef55-4185-8315-a61b5bfc8e22\") " pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.775019 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bba9178-9a01-47ac-9657-c277217019ca-openshift-service-ca\") pod \"perses-operator-5446b9c989-trfdr\" (UID: \"1bba9178-9a01-47ac-9657-c277217019ca\") " pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.775085 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pvtc\" (UniqueName: \"kubernetes.io/projected/1bba9178-9a01-47ac-9657-c277217019ca-kube-api-access-9pvtc\") pod \"perses-operator-5446b9c989-trfdr\" (UID: \"1bba9178-9a01-47ac-9657-c277217019ca\") " pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.784239 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4a067c7-ef55-4185-8315-a61b5bfc8e22-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-bsscx\" (UID: \"f4a067c7-ef55-4185-8315-a61b5bfc8e22\") " pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.789164 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators_194f5401-404a-43b8-b826-3df930390630_0(93f516452936aeaf8f92e5356ff01bdbdb435695bba81500aae63c3f4d12e09d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.789249 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators_194f5401-404a-43b8-b826-3df930390630_0(93f516452936aeaf8f92e5356ff01bdbdb435695bba81500aae63c3f4d12e09d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.789285 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators_194f5401-404a-43b8-b826-3df930390630_0(93f516452936aeaf8f92e5356ff01bdbdb435695bba81500aae63c3f4d12e09d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.789344 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators(194f5401-404a-43b8-b826-3df930390630)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators(194f5401-404a-43b8-b826-3df930390630)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators_194f5401-404a-43b8-b826-3df930390630_0(93f516452936aeaf8f92e5356ff01bdbdb435695bba81500aae63c3f4d12e09d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" podUID="194f5401-404a-43b8-b826-3df930390630" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.796713 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmmx5\" (UniqueName: \"kubernetes.io/projected/f4a067c7-ef55-4185-8315-a61b5bfc8e22-kube-api-access-gmmx5\") pod \"observability-operator-d8bb48f5d-bsscx\" (UID: \"f4a067c7-ef55-4185-8315-a61b5bfc8e22\") " pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.866529 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.884761 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pvtc\" (UniqueName: \"kubernetes.io/projected/1bba9178-9a01-47ac-9657-c277217019ca-kube-api-access-9pvtc\") pod \"perses-operator-5446b9c989-trfdr\" (UID: \"1bba9178-9a01-47ac-9657-c277217019ca\") " pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.884837 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bba9178-9a01-47ac-9657-c277217019ca-openshift-service-ca\") pod \"perses-operator-5446b9c989-trfdr\" (UID: \"1bba9178-9a01-47ac-9657-c277217019ca\") " pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.885771 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/1bba9178-9a01-47ac-9657-c277217019ca-openshift-service-ca\") pod \"perses-operator-5446b9c989-trfdr\" (UID: \"1bba9178-9a01-47ac-9657-c277217019ca\") " pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.889335 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bsscx_openshift-operators_f4a067c7-ef55-4185-8315-a61b5bfc8e22_0(c2fb69881bfded7616d9cc1c5ebf88e07e715a17c0856a9b6453dc93a3d705f7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.889409 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bsscx_openshift-operators_f4a067c7-ef55-4185-8315-a61b5bfc8e22_0(c2fb69881bfded7616d9cc1c5ebf88e07e715a17c0856a9b6453dc93a3d705f7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.889446 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bsscx_openshift-operators_f4a067c7-ef55-4185-8315-a61b5bfc8e22_0(c2fb69881bfded7616d9cc1c5ebf88e07e715a17c0856a9b6453dc93a3d705f7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:47 crc kubenswrapper[4912]: E1203 00:36:47.889522 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-bsscx_openshift-operators(f4a067c7-ef55-4185-8315-a61b5bfc8e22)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-bsscx_openshift-operators(f4a067c7-ef55-4185-8315-a61b5bfc8e22)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bsscx_openshift-operators_f4a067c7-ef55-4185-8315-a61b5bfc8e22_0(c2fb69881bfded7616d9cc1c5ebf88e07e715a17c0856a9b6453dc93a3d705f7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" podUID="f4a067c7-ef55-4185-8315-a61b5bfc8e22" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.905358 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pvtc\" (UniqueName: \"kubernetes.io/projected/1bba9178-9a01-47ac-9657-c277217019ca-kube-api-access-9pvtc\") pod \"perses-operator-5446b9c989-trfdr\" (UID: \"1bba9178-9a01-47ac-9657-c277217019ca\") " pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:47 crc kubenswrapper[4912]: I1203 00:36:47.975706 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.001024 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:48 crc kubenswrapper[4912]: E1203 00:36:48.070938 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-trfdr_openshift-operators_1bba9178-9a01-47ac-9657-c277217019ca_0(0a05874379de9f11b53b0efd4aaa7cd1b15dbdb937d4ed41e1d5afbf41436ae5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:48 crc kubenswrapper[4912]: E1203 00:36:48.071366 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-trfdr_openshift-operators_1bba9178-9a01-47ac-9657-c277217019ca_0(0a05874379de9f11b53b0efd4aaa7cd1b15dbdb937d4ed41e1d5afbf41436ae5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:48 crc kubenswrapper[4912]: E1203 00:36:48.071423 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-trfdr_openshift-operators_1bba9178-9a01-47ac-9657-c277217019ca_0(0a05874379de9f11b53b0efd4aaa7cd1b15dbdb937d4ed41e1d5afbf41436ae5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:48 crc kubenswrapper[4912]: E1203 00:36:48.071558 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-trfdr_openshift-operators(1bba9178-9a01-47ac-9657-c277217019ca)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-trfdr_openshift-operators(1bba9178-9a01-47ac-9657-c277217019ca)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-trfdr_openshift-operators_1bba9178-9a01-47ac-9657-c277217019ca_0(0a05874379de9f11b53b0efd4aaa7cd1b15dbdb937d4ed41e1d5afbf41436ae5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-trfdr" podUID="1bba9178-9a01-47ac-9657-c277217019ca" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.078072 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.078145 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.078200 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.079083 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6db36669257e6a0bbce38d40b2b10188b2e1ca05791ecba752a07c893daa324b"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.079154 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://6db36669257e6a0bbce38d40b2b10188b2e1ca05791ecba752a07c893daa324b" gracePeriod=600 Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.087705 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-catalog-content\") pod \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.087892 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtgz9\" (UniqueName: \"kubernetes.io/projected/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-kube-api-access-vtgz9\") pod \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.087935 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-utilities\") pod \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\" (UID: \"c068c38b-d1c8-4223-b7b6-4da7f113f2ea\") " Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.089455 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-utilities" (OuterVolumeSpecName: "utilities") pod "c068c38b-d1c8-4223-b7b6-4da7f113f2ea" (UID: "c068c38b-d1c8-4223-b7b6-4da7f113f2ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.096631 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-kube-api-access-vtgz9" (OuterVolumeSpecName: "kube-api-access-vtgz9") pod "c068c38b-d1c8-4223-b7b6-4da7f113f2ea" (UID: "c068c38b-d1c8-4223-b7b6-4da7f113f2ea"). InnerVolumeSpecName "kube-api-access-vtgz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.189423 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.189485 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtgz9\" (UniqueName: \"kubernetes.io/projected/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-kube-api-access-vtgz9\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.284003 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c068c38b-d1c8-4223-b7b6-4da7f113f2ea" (UID: "c068c38b-d1c8-4223-b7b6-4da7f113f2ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.290394 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c068c38b-d1c8-4223-b7b6-4da7f113f2ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.675846 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d9mk6" event={"ID":"c068c38b-d1c8-4223-b7b6-4da7f113f2ea","Type":"ContainerDied","Data":"e0423543ca789c6c494698f40c9d1b48499968c5442ed6100887b296aaa65e77"} Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.675897 4912 scope.go:117] "RemoveContainer" containerID="5b901826c74c1f1d6802d6aa2add0af50a30cd07efb6bec7b8f85e41bc3e00d9" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.675960 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d9mk6" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.690830 4912 scope.go:117] "RemoveContainer" containerID="114696955b566cf1731a42db40c00a18e8188b6fbfa076e1104a0586f5681cda" Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.702461 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d9mk6"] Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.706259 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d9mk6"] Dec 03 00:36:48 crc kubenswrapper[4912]: I1203 00:36:48.720190 4912 scope.go:117] "RemoveContainer" containerID="05e153b239407ed28bfa1130bfc504740a0782c86e25016e30a20c322cd18456" Dec 03 00:36:49 crc kubenswrapper[4912]: I1203 00:36:49.685823 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="6db36669257e6a0bbce38d40b2b10188b2e1ca05791ecba752a07c893daa324b" exitCode=0 Dec 03 00:36:49 crc kubenswrapper[4912]: I1203 00:36:49.685875 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"6db36669257e6a0bbce38d40b2b10188b2e1ca05791ecba752a07c893daa324b"} Dec 03 00:36:49 crc kubenswrapper[4912]: I1203 00:36:49.686200 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"8ef572750b10ee1c03bff18588f184d2f43af325eeff79f4d4e863d818a97628"} Dec 03 00:36:49 crc kubenswrapper[4912]: I1203 00:36:49.686226 4912 scope.go:117] "RemoveContainer" containerID="67d01fbca4f2a6414cad35bc8cfd7bf8c1cdf2df989831ac2d2a99ac36bfb004" Dec 03 00:36:49 crc kubenswrapper[4912]: I1203 00:36:49.695693 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" event={"ID":"2b7ed935-b2f4-470a-ad77-4a2db9011ca6","Type":"ContainerStarted","Data":"3167f2b349fb92c2b84e1ba4c06239141b0cef4820bb59c6a80a494a0f08af3a"} Dec 03 00:36:49 crc kubenswrapper[4912]: I1203 00:36:49.696036 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:49 crc kubenswrapper[4912]: I1203 00:36:49.733196 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:49 crc kubenswrapper[4912]: I1203 00:36:49.772985 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" podStartSLOduration=8.772967571 podStartE2EDuration="8.772967571s" podCreationTimestamp="2025-12-03 00:36:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:36:49.772537199 +0000 UTC m=+795.414557779" watchObservedRunningTime="2025-12-03 00:36:49.772967571 +0000 UTC m=+795.414988131" Dec 03 00:36:50 crc kubenswrapper[4912]: I1203 00:36:50.710043 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" path="/var/lib/kubelet/pods/c068c38b-d1c8-4223-b7b6-4da7f113f2ea/volumes" Dec 03 00:36:50 crc kubenswrapper[4912]: I1203 00:36:50.711191 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:50 crc kubenswrapper[4912]: I1203 00:36:50.711997 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:50 crc kubenswrapper[4912]: I1203 00:36:50.812114 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.597042 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5"] Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.597263 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.597916 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.616016 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8"] Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.616147 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.616608 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.643593 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-trfdr"] Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.643767 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.644334 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.659654 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2"] Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.659787 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.660280 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.684128 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-bsscx"] Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.684250 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:51 crc kubenswrapper[4912]: I1203 00:36:51.690988 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.740624 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators_194f5401-404a-43b8-b826-3df930390630_0(4b8281b484a6bc40d26835d235cd949ca2cd4380a0e33a240f55510107511522): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.740738 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators_194f5401-404a-43b8-b826-3df930390630_0(4b8281b484a6bc40d26835d235cd949ca2cd4380a0e33a240f55510107511522): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.740766 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators_194f5401-404a-43b8-b826-3df930390630_0(4b8281b484a6bc40d26835d235cd949ca2cd4380a0e33a240f55510107511522): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.740899 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators(194f5401-404a-43b8-b826-3df930390630)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators(194f5401-404a-43b8-b826-3df930390630)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_openshift-operators_194f5401-404a-43b8-b826-3df930390630_0(4b8281b484a6bc40d26835d235cd949ca2cd4380a0e33a240f55510107511522): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" podUID="194f5401-404a-43b8-b826-3df930390630" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.769728 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators_377b37d8-070e-4463-9ea8-c698c4e289a4_0(8ce362ffeb7aecddc64e086953d11521b3e60641132818b1b150a1258be58941): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.769801 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators_377b37d8-070e-4463-9ea8-c698c4e289a4_0(8ce362ffeb7aecddc64e086953d11521b3e60641132818b1b150a1258be58941): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.769828 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators_377b37d8-070e-4463-9ea8-c698c4e289a4_0(8ce362ffeb7aecddc64e086953d11521b3e60641132818b1b150a1258be58941): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.769867 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators(377b37d8-070e-4463-9ea8-c698c4e289a4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators(377b37d8-070e-4463-9ea8-c698c4e289a4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-spkm8_openshift-operators_377b37d8-070e-4463-9ea8-c698c4e289a4_0(8ce362ffeb7aecddc64e086953d11521b3e60641132818b1b150a1258be58941): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" podUID="377b37d8-070e-4463-9ea8-c698c4e289a4" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.803697 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-trfdr_openshift-operators_1bba9178-9a01-47ac-9657-c277217019ca_0(1ae267b744f680bd2bc2790222095164504378b58e4d44a54325a44339f52933): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.803808 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-trfdr_openshift-operators_1bba9178-9a01-47ac-9657-c277217019ca_0(1ae267b744f680bd2bc2790222095164504378b58e4d44a54325a44339f52933): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.803852 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-trfdr_openshift-operators_1bba9178-9a01-47ac-9657-c277217019ca_0(1ae267b744f680bd2bc2790222095164504378b58e4d44a54325a44339f52933): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.803924 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-trfdr_openshift-operators(1bba9178-9a01-47ac-9657-c277217019ca)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-trfdr_openshift-operators(1bba9178-9a01-47ac-9657-c277217019ca)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-trfdr_openshift-operators_1bba9178-9a01-47ac-9657-c277217019ca_0(1ae267b744f680bd2bc2790222095164504378b58e4d44a54325a44339f52933): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-trfdr" podUID="1bba9178-9a01-47ac-9657-c277217019ca" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.833961 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0_0(421996a48d263e436da790bec9525e52f90bf00015195607444efa372b09ee88): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.834071 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0_0(421996a48d263e436da790bec9525e52f90bf00015195607444efa372b09ee88): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.834103 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0_0(421996a48d263e436da790bec9525e52f90bf00015195607444efa372b09ee88): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.834171 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators(45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators(45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_openshift-operators_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0_0(421996a48d263e436da790bec9525e52f90bf00015195607444efa372b09ee88): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" podUID="45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.843815 4912 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bsscx_openshift-operators_f4a067c7-ef55-4185-8315-a61b5bfc8e22_0(f36b581e24293aa1846f7c9de0682c65124b42c1d7c8b07623c449062c8e355a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.843887 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bsscx_openshift-operators_f4a067c7-ef55-4185-8315-a61b5bfc8e22_0(f36b581e24293aa1846f7c9de0682c65124b42c1d7c8b07623c449062c8e355a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.843907 4912 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bsscx_openshift-operators_f4a067c7-ef55-4185-8315-a61b5bfc8e22_0(f36b581e24293aa1846f7c9de0682c65124b42c1d7c8b07623c449062c8e355a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:36:51 crc kubenswrapper[4912]: E1203 00:36:51.843950 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-bsscx_openshift-operators(f4a067c7-ef55-4185-8315-a61b5bfc8e22)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-bsscx_openshift-operators(f4a067c7-ef55-4185-8315-a61b5bfc8e22)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-bsscx_openshift-operators_f4a067c7-ef55-4185-8315-a61b5bfc8e22_0(f36b581e24293aa1846f7c9de0682c65124b42c1d7c8b07623c449062c8e355a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" podUID="f4a067c7-ef55-4185-8315-a61b5bfc8e22" Dec 03 00:37:02 crc kubenswrapper[4912]: I1203 00:37:02.570579 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:37:02 crc kubenswrapper[4912]: I1203 00:37:02.571549 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:37:02 crc kubenswrapper[4912]: I1203 00:37:02.939576 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-trfdr"] Dec 03 00:37:02 crc kubenswrapper[4912]: W1203 00:37:02.951264 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bba9178_9a01_47ac_9657_c277217019ca.slice/crio-fb86ee4cb721e52caa4bcc04b79013e988902b0f73b7f152495324ab4879e09b WatchSource:0}: Error finding container fb86ee4cb721e52caa4bcc04b79013e988902b0f73b7f152495324ab4879e09b: Status 404 returned error can't find the container with id fb86ee4cb721e52caa4bcc04b79013e988902b0f73b7f152495324ab4879e09b Dec 03 00:37:03 crc kubenswrapper[4912]: I1203 00:37:03.820265 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-trfdr" event={"ID":"1bba9178-9a01-47ac-9657-c277217019ca","Type":"ContainerStarted","Data":"fb86ee4cb721e52caa4bcc04b79013e988902b0f73b7f152495324ab4879e09b"} Dec 03 00:37:05 crc kubenswrapper[4912]: I1203 00:37:05.574854 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:37:05 crc kubenswrapper[4912]: I1203 00:37:05.575017 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:37:05 crc kubenswrapper[4912]: I1203 00:37:05.575689 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:37:05 crc kubenswrapper[4912]: I1203 00:37:05.576515 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" Dec 03 00:37:05 crc kubenswrapper[4912]: I1203 00:37:05.834356 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-bsscx"] Dec 03 00:37:05 crc kubenswrapper[4912]: I1203 00:37:05.912157 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5"] Dec 03 00:37:05 crc kubenswrapper[4912]: W1203 00:37:05.923144 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod194f5401_404a_43b8_b826_3df930390630.slice/crio-bd49b67e9c7ccf8f22196adab002cdb5506377718a305a9da377f3f27c09b45f WatchSource:0}: Error finding container bd49b67e9c7ccf8f22196adab002cdb5506377718a305a9da377f3f27c09b45f: Status 404 returned error can't find the container with id bd49b67e9c7ccf8f22196adab002cdb5506377718a305a9da377f3f27c09b45f Dec 03 00:37:06 crc kubenswrapper[4912]: I1203 00:37:06.571578 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:37:06 crc kubenswrapper[4912]: I1203 00:37:06.572023 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" Dec 03 00:37:06 crc kubenswrapper[4912]: I1203 00:37:06.869579 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" event={"ID":"f4a067c7-ef55-4185-8315-a61b5bfc8e22","Type":"ContainerStarted","Data":"c6e552a05eb5abb805e28a028eb86ccbdb1997f490e13197f912217ff1912e1e"} Dec 03 00:37:06 crc kubenswrapper[4912]: I1203 00:37:06.884614 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" event={"ID":"194f5401-404a-43b8-b826-3df930390630","Type":"ContainerStarted","Data":"bd49b67e9c7ccf8f22196adab002cdb5506377718a305a9da377f3f27c09b45f"} Dec 03 00:37:07 crc kubenswrapper[4912]: I1203 00:37:07.025680 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8"] Dec 03 00:37:07 crc kubenswrapper[4912]: I1203 00:37:07.571516 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:37:07 crc kubenswrapper[4912]: I1203 00:37:07.572084 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" Dec 03 00:37:11 crc kubenswrapper[4912]: I1203 00:37:11.914608 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" event={"ID":"377b37d8-070e-4463-9ea8-c698c4e289a4","Type":"ContainerStarted","Data":"3e919b0e097643590ed7c01a88900e9f65bfd8c51b3a0cd575aa615923e7931e"} Dec 03 00:37:11 crc kubenswrapper[4912]: I1203 00:37:11.942957 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hcxn8" Dec 03 00:37:12 crc kubenswrapper[4912]: I1203 00:37:12.619358 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2"] Dec 03 00:37:13 crc kubenswrapper[4912]: W1203 00:37:13.697254 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45c9685c_8c5e_4da5_9c6f_d1f00f6d83e0.slice/crio-07b0d38a7067b14e1fd5a007d26d0b28071e3a0ee5623d7c4043cd465baf2471 WatchSource:0}: Error finding container 07b0d38a7067b14e1fd5a007d26d0b28071e3a0ee5623d7c4043cd465baf2471: Status 404 returned error can't find the container with id 07b0d38a7067b14e1fd5a007d26d0b28071e3a0ee5623d7c4043cd465baf2471 Dec 03 00:37:13 crc kubenswrapper[4912]: I1203 00:37:13.931639 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-trfdr" event={"ID":"1bba9178-9a01-47ac-9657-c277217019ca","Type":"ContainerStarted","Data":"ced3a19832b054c5e273cf8bf043b2f4f54eafd39d3a0a4dd3326a01d7d482a4"} Dec 03 00:37:13 crc kubenswrapper[4912]: I1203 00:37:13.932018 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:37:13 crc kubenswrapper[4912]: I1203 00:37:13.936642 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" event={"ID":"194f5401-404a-43b8-b826-3df930390630","Type":"ContainerStarted","Data":"f65f74b37ff375b07c6535086d027fbf9d2a36e246ee994584c37aba0456dcbe"} Dec 03 00:37:13 crc kubenswrapper[4912]: I1203 00:37:13.938579 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" event={"ID":"45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0","Type":"ContainerStarted","Data":"07b0d38a7067b14e1fd5a007d26d0b28071e3a0ee5623d7c4043cd465baf2471"} Dec 03 00:37:13 crc kubenswrapper[4912]: I1203 00:37:13.952374 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-trfdr" podStartSLOduration=16.163652412 podStartE2EDuration="26.952352745s" podCreationTimestamp="2025-12-03 00:36:47 +0000 UTC" firstStartedPulling="2025-12-03 00:37:02.954453614 +0000 UTC m=+808.596474174" lastFinishedPulling="2025-12-03 00:37:13.743153947 +0000 UTC m=+819.385174507" observedRunningTime="2025-12-03 00:37:13.950831584 +0000 UTC m=+819.592852164" watchObservedRunningTime="2025-12-03 00:37:13.952352745 +0000 UTC m=+819.594373305" Dec 03 00:37:13 crc kubenswrapper[4912]: I1203 00:37:13.983010 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5" podStartSLOduration=19.151806825 podStartE2EDuration="26.982981397s" podCreationTimestamp="2025-12-03 00:36:47 +0000 UTC" firstStartedPulling="2025-12-03 00:37:05.927495766 +0000 UTC m=+811.569516326" lastFinishedPulling="2025-12-03 00:37:13.758670338 +0000 UTC m=+819.400690898" observedRunningTime="2025-12-03 00:37:13.978318181 +0000 UTC m=+819.620338741" watchObservedRunningTime="2025-12-03 00:37:13.982981397 +0000 UTC m=+819.625001967" Dec 03 00:37:18 crc kubenswrapper[4912]: I1203 00:37:18.004386 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-trfdr" Dec 03 00:37:19 crc kubenswrapper[4912]: I1203 00:37:19.985335 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" event={"ID":"45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0","Type":"ContainerStarted","Data":"6338359169bd5517f806f10aa68e65078af75d7c2496b2f92efb271d8ab7fa41"} Dec 03 00:37:19 crc kubenswrapper[4912]: I1203 00:37:19.987002 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" event={"ID":"377b37d8-070e-4463-9ea8-c698c4e289a4","Type":"ContainerStarted","Data":"7f2703462c21522620a89576f56651fa81ed0cb600169547b8e11a82b71bbaf3"} Dec 03 00:37:19 crc kubenswrapper[4912]: I1203 00:37:19.988750 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" event={"ID":"f4a067c7-ef55-4185-8315-a61b5bfc8e22","Type":"ContainerStarted","Data":"07aa928beaee3271857d7a96ec9cc6227be71340ef37848532e920faffd3452a"} Dec 03 00:37:19 crc kubenswrapper[4912]: I1203 00:37:19.988975 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:37:20 crc kubenswrapper[4912]: I1203 00:37:20.005684 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2" podStartSLOduration=27.175719527 podStartE2EDuration="33.005665297s" podCreationTimestamp="2025-12-03 00:36:47 +0000 UTC" firstStartedPulling="2025-12-03 00:37:13.702074031 +0000 UTC m=+819.344094591" lastFinishedPulling="2025-12-03 00:37:19.532019801 +0000 UTC m=+825.174040361" observedRunningTime="2025-12-03 00:37:20.003228352 +0000 UTC m=+825.645248922" watchObservedRunningTime="2025-12-03 00:37:20.005665297 +0000 UTC m=+825.647685857" Dec 03 00:37:20 crc kubenswrapper[4912]: I1203 00:37:20.039052 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" podStartSLOduration=19.352724437 podStartE2EDuration="33.039034113s" podCreationTimestamp="2025-12-03 00:36:47 +0000 UTC" firstStartedPulling="2025-12-03 00:37:05.869545212 +0000 UTC m=+811.511565772" lastFinishedPulling="2025-12-03 00:37:19.555854888 +0000 UTC m=+825.197875448" observedRunningTime="2025-12-03 00:37:20.033821351 +0000 UTC m=+825.675841931" watchObservedRunningTime="2025-12-03 00:37:20.039034113 +0000 UTC m=+825.681054673" Dec 03 00:37:20 crc kubenswrapper[4912]: I1203 00:37:20.062600 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-spkm8" podStartSLOduration=24.82503535 podStartE2EDuration="33.062577492s" podCreationTimestamp="2025-12-03 00:36:47 +0000 UTC" firstStartedPulling="2025-12-03 00:37:11.311568743 +0000 UTC m=+816.953589303" lastFinishedPulling="2025-12-03 00:37:19.549110885 +0000 UTC m=+825.191131445" observedRunningTime="2025-12-03 00:37:20.056647692 +0000 UTC m=+825.698668272" watchObservedRunningTime="2025-12-03 00:37:20.062577492 +0000 UTC m=+825.704598062" Dec 03 00:37:20 crc kubenswrapper[4912]: I1203 00:37:20.090057 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-bsscx" Dec 03 00:37:27 crc kubenswrapper[4912]: I1203 00:37:27.997496 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dlkbd"] Dec 03 00:37:27 crc kubenswrapper[4912]: E1203 00:37:27.998317 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerName="registry-server" Dec 03 00:37:27 crc kubenswrapper[4912]: I1203 00:37:27.998337 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerName="registry-server" Dec 03 00:37:27 crc kubenswrapper[4912]: E1203 00:37:27.998350 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerName="extract-utilities" Dec 03 00:37:27 crc kubenswrapper[4912]: I1203 00:37:27.998359 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerName="extract-utilities" Dec 03 00:37:27 crc kubenswrapper[4912]: E1203 00:37:27.998372 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerName="extract-content" Dec 03 00:37:27 crc kubenswrapper[4912]: I1203 00:37:27.998378 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerName="extract-content" Dec 03 00:37:27 crc kubenswrapper[4912]: I1203 00:37:27.998525 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c068c38b-d1c8-4223-b7b6-4da7f113f2ea" containerName="registry-server" Dec 03 00:37:27 crc kubenswrapper[4912]: I1203 00:37:27.999460 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-dlkbd" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.004125 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.004126 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.004185 4912 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-bgsxl" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.008083 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9hkxj"] Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.008994 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9hkxj" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.010443 4912 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7jd4t" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.014305 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dlkbd"] Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.034243 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-s6b5l"] Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.035348 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.037685 4912 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-4gh42" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.040871 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9hkxj"] Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.053737 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46s2d\" (UniqueName: \"kubernetes.io/projected/2620530c-1377-4da3-a983-a0211b396c18-kube-api-access-46s2d\") pod \"cert-manager-cainjector-7f985d654d-dlkbd\" (UID: \"2620530c-1377-4da3-a983-a0211b396c18\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dlkbd" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.053781 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gm7q\" (UniqueName: \"kubernetes.io/projected/77d9d66c-7ec2-4cc0-8240-bd02f4b710c7-kube-api-access-8gm7q\") pod \"cert-manager-webhook-5655c58dd6-s6b5l\" (UID: \"77d9d66c-7ec2-4cc0-8240-bd02f4b710c7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.053914 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpb8s\" (UniqueName: \"kubernetes.io/projected/52072628-05d2-4d43-8fea-11ec571a1a5f-kube-api-access-qpb8s\") pod \"cert-manager-5b446d88c5-9hkxj\" (UID: \"52072628-05d2-4d43-8fea-11ec571a1a5f\") " pod="cert-manager/cert-manager-5b446d88c5-9hkxj" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.057890 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-s6b5l"] Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.155882 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46s2d\" (UniqueName: \"kubernetes.io/projected/2620530c-1377-4da3-a983-a0211b396c18-kube-api-access-46s2d\") pod \"cert-manager-cainjector-7f985d654d-dlkbd\" (UID: \"2620530c-1377-4da3-a983-a0211b396c18\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dlkbd" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.155995 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gm7q\" (UniqueName: \"kubernetes.io/projected/77d9d66c-7ec2-4cc0-8240-bd02f4b710c7-kube-api-access-8gm7q\") pod \"cert-manager-webhook-5655c58dd6-s6b5l\" (UID: \"77d9d66c-7ec2-4cc0-8240-bd02f4b710c7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.156080 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpb8s\" (UniqueName: \"kubernetes.io/projected/52072628-05d2-4d43-8fea-11ec571a1a5f-kube-api-access-qpb8s\") pod \"cert-manager-5b446d88c5-9hkxj\" (UID: \"52072628-05d2-4d43-8fea-11ec571a1a5f\") " pod="cert-manager/cert-manager-5b446d88c5-9hkxj" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.180678 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gm7q\" (UniqueName: \"kubernetes.io/projected/77d9d66c-7ec2-4cc0-8240-bd02f4b710c7-kube-api-access-8gm7q\") pod \"cert-manager-webhook-5655c58dd6-s6b5l\" (UID: \"77d9d66c-7ec2-4cc0-8240-bd02f4b710c7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.180690 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpb8s\" (UniqueName: \"kubernetes.io/projected/52072628-05d2-4d43-8fea-11ec571a1a5f-kube-api-access-qpb8s\") pod \"cert-manager-5b446d88c5-9hkxj\" (UID: \"52072628-05d2-4d43-8fea-11ec571a1a5f\") " pod="cert-manager/cert-manager-5b446d88c5-9hkxj" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.196478 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46s2d\" (UniqueName: \"kubernetes.io/projected/2620530c-1377-4da3-a983-a0211b396c18-kube-api-access-46s2d\") pod \"cert-manager-cainjector-7f985d654d-dlkbd\" (UID: \"2620530c-1377-4da3-a983-a0211b396c18\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-dlkbd" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.320920 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-dlkbd" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.329149 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-9hkxj" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.352391 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.892786 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-9hkxj"] Dec 03 00:37:28 crc kubenswrapper[4912]: W1203 00:37:28.898743 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52072628_05d2_4d43_8fea_11ec571a1a5f.slice/crio-5a9a9708942c40122209906dccdf81d13fad7d0c6e950ab2e2c1d36356572a28 WatchSource:0}: Error finding container 5a9a9708942c40122209906dccdf81d13fad7d0c6e950ab2e2c1d36356572a28: Status 404 returned error can't find the container with id 5a9a9708942c40122209906dccdf81d13fad7d0c6e950ab2e2c1d36356572a28 Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.908624 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-s6b5l"] Dec 03 00:37:28 crc kubenswrapper[4912]: W1203 00:37:28.914406 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77d9d66c_7ec2_4cc0_8240_bd02f4b710c7.slice/crio-0e1c8e20a61ab0a3b9b91318be0f620ece26d30b8a0be722c8410d279a2de760 WatchSource:0}: Error finding container 0e1c8e20a61ab0a3b9b91318be0f620ece26d30b8a0be722c8410d279a2de760: Status 404 returned error can't find the container with id 0e1c8e20a61ab0a3b9b91318be0f620ece26d30b8a0be722c8410d279a2de760 Dec 03 00:37:28 crc kubenswrapper[4912]: I1203 00:37:28.926705 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-dlkbd"] Dec 03 00:37:28 crc kubenswrapper[4912]: W1203 00:37:28.930022 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2620530c_1377_4da3_a983_a0211b396c18.slice/crio-d0899066baa212fcee4a90c80a1b0db3510832e27d87179fc724f8f4d6f88734 WatchSource:0}: Error finding container d0899066baa212fcee4a90c80a1b0db3510832e27d87179fc724f8f4d6f88734: Status 404 returned error can't find the container with id d0899066baa212fcee4a90c80a1b0db3510832e27d87179fc724f8f4d6f88734 Dec 03 00:37:29 crc kubenswrapper[4912]: I1203 00:37:29.037540 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-dlkbd" event={"ID":"2620530c-1377-4da3-a983-a0211b396c18","Type":"ContainerStarted","Data":"d0899066baa212fcee4a90c80a1b0db3510832e27d87179fc724f8f4d6f88734"} Dec 03 00:37:29 crc kubenswrapper[4912]: I1203 00:37:29.038964 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9hkxj" event={"ID":"52072628-05d2-4d43-8fea-11ec571a1a5f","Type":"ContainerStarted","Data":"5a9a9708942c40122209906dccdf81d13fad7d0c6e950ab2e2c1d36356572a28"} Dec 03 00:37:29 crc kubenswrapper[4912]: I1203 00:37:29.040592 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" event={"ID":"77d9d66c-7ec2-4cc0-8240-bd02f4b710c7","Type":"ContainerStarted","Data":"0e1c8e20a61ab0a3b9b91318be0f620ece26d30b8a0be722c8410d279a2de760"} Dec 03 00:37:34 crc kubenswrapper[4912]: I1203 00:37:34.075343 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-9hkxj" event={"ID":"52072628-05d2-4d43-8fea-11ec571a1a5f","Type":"ContainerStarted","Data":"8616e1580ab08efec8d38d71aed5b3efc65ec9f0006e6950df8350fa29a600c4"} Dec 03 00:37:34 crc kubenswrapper[4912]: I1203 00:37:34.077567 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" event={"ID":"77d9d66c-7ec2-4cc0-8240-bd02f4b710c7","Type":"ContainerStarted","Data":"429c317c4891bdec42cf33b19f9f89bb2d06dc7be67fca694903846d1c6dc30a"} Dec 03 00:37:34 crc kubenswrapper[4912]: I1203 00:37:34.077656 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" Dec 03 00:37:34 crc kubenswrapper[4912]: I1203 00:37:34.079057 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-dlkbd" event={"ID":"2620530c-1377-4da3-a983-a0211b396c18","Type":"ContainerStarted","Data":"18373c67c9f0e46952517bb48c9b38cea29cc9df181bcf5e813ca084a70856ff"} Dec 03 00:37:34 crc kubenswrapper[4912]: I1203 00:37:34.096619 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-9hkxj" podStartSLOduration=3.091003706 podStartE2EDuration="7.096596436s" podCreationTimestamp="2025-12-03 00:37:27 +0000 UTC" firstStartedPulling="2025-12-03 00:37:28.901252202 +0000 UTC m=+834.543272762" lastFinishedPulling="2025-12-03 00:37:32.906844932 +0000 UTC m=+838.548865492" observedRunningTime="2025-12-03 00:37:34.095304991 +0000 UTC m=+839.737325571" watchObservedRunningTime="2025-12-03 00:37:34.096596436 +0000 UTC m=+839.738617006" Dec 03 00:37:34 crc kubenswrapper[4912]: I1203 00:37:34.160729 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-dlkbd" podStartSLOduration=3.2596056239999998 podStartE2EDuration="7.160703977s" podCreationTimestamp="2025-12-03 00:37:27 +0000 UTC" firstStartedPulling="2025-12-03 00:37:28.932357006 +0000 UTC m=+834.574377566" lastFinishedPulling="2025-12-03 00:37:32.833455369 +0000 UTC m=+838.475475919" observedRunningTime="2025-12-03 00:37:34.159540195 +0000 UTC m=+839.801560775" watchObservedRunningTime="2025-12-03 00:37:34.160703977 +0000 UTC m=+839.802724547" Dec 03 00:37:34 crc kubenswrapper[4912]: I1203 00:37:34.191276 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" podStartSLOduration=2.27493203 podStartE2EDuration="6.191246896s" podCreationTimestamp="2025-12-03 00:37:28 +0000 UTC" firstStartedPulling="2025-12-03 00:37:28.916528217 +0000 UTC m=+834.558548777" lastFinishedPulling="2025-12-03 00:37:32.832843083 +0000 UTC m=+838.474863643" observedRunningTime="2025-12-03 00:37:34.181622514 +0000 UTC m=+839.823643074" watchObservedRunningTime="2025-12-03 00:37:34.191246896 +0000 UTC m=+839.833267456" Dec 03 00:37:38 crc kubenswrapper[4912]: I1203 00:37:38.355563 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-s6b5l" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.165033 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf"] Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.166693 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.175166 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.183734 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf"] Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.283879 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.284013 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.284056 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj89j\" (UniqueName: \"kubernetes.io/projected/26654d73-149b-418b-8c5e-00653837545b-kube-api-access-bj89j\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.385502 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.385571 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj89j\" (UniqueName: \"kubernetes.io/projected/26654d73-149b-418b-8c5e-00653837545b-kube-api-access-bj89j\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.385630 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.386499 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.386649 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.411928 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj89j\" (UniqueName: \"kubernetes.io/projected/26654d73-149b-418b-8c5e-00653837545b-kube-api-access-bj89j\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.485768 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.568950 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh"] Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.581197 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.593408 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh"] Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.691070 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.691203 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrhpr\" (UniqueName: \"kubernetes.io/projected/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-kube-api-access-wrhpr\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.691242 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.721489 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf"] Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.793168 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrhpr\" (UniqueName: \"kubernetes.io/projected/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-kube-api-access-wrhpr\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.793739 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.793826 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.794349 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.794482 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.813535 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrhpr\" (UniqueName: \"kubernetes.io/projected/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-kube-api-access-wrhpr\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:09 crc kubenswrapper[4912]: I1203 00:38:09.913553 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:10 crc kubenswrapper[4912]: I1203 00:38:10.136314 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh"] Dec 03 00:38:10 crc kubenswrapper[4912]: W1203 00:38:10.139421 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25cacc35_89f1_4f0e_a621_ee002ce1f7e0.slice/crio-e81015c0f5bbb5df99f28f6b91bbb221ffe9896190b645e0c8ffa2522432f620 WatchSource:0}: Error finding container e81015c0f5bbb5df99f28f6b91bbb221ffe9896190b645e0c8ffa2522432f620: Status 404 returned error can't find the container with id e81015c0f5bbb5df99f28f6b91bbb221ffe9896190b645e0c8ffa2522432f620 Dec 03 00:38:10 crc kubenswrapper[4912]: I1203 00:38:10.307253 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" event={"ID":"25cacc35-89f1-4f0e-a621-ee002ce1f7e0","Type":"ContainerStarted","Data":"a625959d0b0c7dcfac51ca20cfad7558527debafa566ea7c464209859bb47574"} Dec 03 00:38:10 crc kubenswrapper[4912]: I1203 00:38:10.307316 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" event={"ID":"25cacc35-89f1-4f0e-a621-ee002ce1f7e0","Type":"ContainerStarted","Data":"e81015c0f5bbb5df99f28f6b91bbb221ffe9896190b645e0c8ffa2522432f620"} Dec 03 00:38:10 crc kubenswrapper[4912]: I1203 00:38:10.309642 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" event={"ID":"26654d73-149b-418b-8c5e-00653837545b","Type":"ContainerStarted","Data":"828f58697f68a1cd99ead5b3f4da3eeb98877cc8fdf9ba4b60e2c7076dd7116d"} Dec 03 00:38:10 crc kubenswrapper[4912]: I1203 00:38:10.309708 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" event={"ID":"26654d73-149b-418b-8c5e-00653837545b","Type":"ContainerStarted","Data":"b688533cd8d84fe35a72485dc4610d89aebe41523c9a9792bdb75bfed35d1886"} Dec 03 00:38:11 crc kubenswrapper[4912]: I1203 00:38:11.318676 4912 generic.go:334] "Generic (PLEG): container finished" podID="26654d73-149b-418b-8c5e-00653837545b" containerID="828f58697f68a1cd99ead5b3f4da3eeb98877cc8fdf9ba4b60e2c7076dd7116d" exitCode=0 Dec 03 00:38:11 crc kubenswrapper[4912]: I1203 00:38:11.318730 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" event={"ID":"26654d73-149b-418b-8c5e-00653837545b","Type":"ContainerDied","Data":"828f58697f68a1cd99ead5b3f4da3eeb98877cc8fdf9ba4b60e2c7076dd7116d"} Dec 03 00:38:11 crc kubenswrapper[4912]: I1203 00:38:11.322767 4912 generic.go:334] "Generic (PLEG): container finished" podID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerID="a625959d0b0c7dcfac51ca20cfad7558527debafa566ea7c464209859bb47574" exitCode=0 Dec 03 00:38:11 crc kubenswrapper[4912]: I1203 00:38:11.322836 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" event={"ID":"25cacc35-89f1-4f0e-a621-ee002ce1f7e0","Type":"ContainerDied","Data":"a625959d0b0c7dcfac51ca20cfad7558527debafa566ea7c464209859bb47574"} Dec 03 00:38:13 crc kubenswrapper[4912]: I1203 00:38:13.337403 4912 generic.go:334] "Generic (PLEG): container finished" podID="26654d73-149b-418b-8c5e-00653837545b" containerID="0a6e15ece03c0eab57449fac9717d8d0c789a31c60a6f0a963fbb0ab9886703a" exitCode=0 Dec 03 00:38:13 crc kubenswrapper[4912]: I1203 00:38:13.337508 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" event={"ID":"26654d73-149b-418b-8c5e-00653837545b","Type":"ContainerDied","Data":"0a6e15ece03c0eab57449fac9717d8d0c789a31c60a6f0a963fbb0ab9886703a"} Dec 03 00:38:14 crc kubenswrapper[4912]: I1203 00:38:14.346248 4912 generic.go:334] "Generic (PLEG): container finished" podID="26654d73-149b-418b-8c5e-00653837545b" containerID="7483b5e24fa1c0fce2aff13cfadd45aacaa6301c544748a8ce55cb2c5cdd8c1e" exitCode=0 Dec 03 00:38:14 crc kubenswrapper[4912]: I1203 00:38:14.346302 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" event={"ID":"26654d73-149b-418b-8c5e-00653837545b","Type":"ContainerDied","Data":"7483b5e24fa1c0fce2aff13cfadd45aacaa6301c544748a8ce55cb2c5cdd8c1e"} Dec 03 00:38:14 crc kubenswrapper[4912]: I1203 00:38:14.348819 4912 generic.go:334] "Generic (PLEG): container finished" podID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerID="5040f053b5e88fb33a048dc51ee5c5aa138f9dca136237ea5bf9f8014a0c0149" exitCode=0 Dec 03 00:38:14 crc kubenswrapper[4912]: I1203 00:38:14.348867 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" event={"ID":"25cacc35-89f1-4f0e-a621-ee002ce1f7e0","Type":"ContainerDied","Data":"5040f053b5e88fb33a048dc51ee5c5aa138f9dca136237ea5bf9f8014a0c0149"} Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.358530 4912 generic.go:334] "Generic (PLEG): container finished" podID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerID="694d0062c32716ce9f0a4a378229abb96a5e5af3e1ce6a875903ee3bee9646a7" exitCode=0 Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.358605 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" event={"ID":"25cacc35-89f1-4f0e-a621-ee002ce1f7e0","Type":"ContainerDied","Data":"694d0062c32716ce9f0a4a378229abb96a5e5af3e1ce6a875903ee3bee9646a7"} Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.678832 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.784492 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-bundle\") pod \"26654d73-149b-418b-8c5e-00653837545b\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.784610 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj89j\" (UniqueName: \"kubernetes.io/projected/26654d73-149b-418b-8c5e-00653837545b-kube-api-access-bj89j\") pod \"26654d73-149b-418b-8c5e-00653837545b\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.784706 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-util\") pod \"26654d73-149b-418b-8c5e-00653837545b\" (UID: \"26654d73-149b-418b-8c5e-00653837545b\") " Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.785582 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-bundle" (OuterVolumeSpecName: "bundle") pod "26654d73-149b-418b-8c5e-00653837545b" (UID: "26654d73-149b-418b-8c5e-00653837545b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.790597 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26654d73-149b-418b-8c5e-00653837545b-kube-api-access-bj89j" (OuterVolumeSpecName: "kube-api-access-bj89j") pod "26654d73-149b-418b-8c5e-00653837545b" (UID: "26654d73-149b-418b-8c5e-00653837545b"). InnerVolumeSpecName "kube-api-access-bj89j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.886382 4912 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:38:15 crc kubenswrapper[4912]: I1203 00:38:15.886460 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj89j\" (UniqueName: \"kubernetes.io/projected/26654d73-149b-418b-8c5e-00653837545b-kube-api-access-bj89j\") on node \"crc\" DevicePath \"\"" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.042300 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-util" (OuterVolumeSpecName: "util") pod "26654d73-149b-418b-8c5e-00653837545b" (UID: "26654d73-149b-418b-8c5e-00653837545b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.088794 4912 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26654d73-149b-418b-8c5e-00653837545b-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.368166 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" event={"ID":"26654d73-149b-418b-8c5e-00653837545b","Type":"ContainerDied","Data":"b688533cd8d84fe35a72485dc4610d89aebe41523c9a9792bdb75bfed35d1886"} Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.368207 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b688533cd8d84fe35a72485dc4610d89aebe41523c9a9792bdb75bfed35d1886" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.368301 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.572291 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.696606 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-util\") pod \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.696928 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-bundle\") pod \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.696952 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrhpr\" (UniqueName: \"kubernetes.io/projected/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-kube-api-access-wrhpr\") pod \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\" (UID: \"25cacc35-89f1-4f0e-a621-ee002ce1f7e0\") " Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.698309 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-bundle" (OuterVolumeSpecName: "bundle") pod "25cacc35-89f1-4f0e-a621-ee002ce1f7e0" (UID: "25cacc35-89f1-4f0e-a621-ee002ce1f7e0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.704354 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-kube-api-access-wrhpr" (OuterVolumeSpecName: "kube-api-access-wrhpr") pod "25cacc35-89f1-4f0e-a621-ee002ce1f7e0" (UID: "25cacc35-89f1-4f0e-a621-ee002ce1f7e0"). InnerVolumeSpecName "kube-api-access-wrhpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.707325 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-util" (OuterVolumeSpecName: "util") pod "25cacc35-89f1-4f0e-a621-ee002ce1f7e0" (UID: "25cacc35-89f1-4f0e-a621-ee002ce1f7e0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.799371 4912 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.799449 4912 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:38:16 crc kubenswrapper[4912]: I1203 00:38:16.799464 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrhpr\" (UniqueName: \"kubernetes.io/projected/25cacc35-89f1-4f0e-a621-ee002ce1f7e0-kube-api-access-wrhpr\") on node \"crc\" DevicePath \"\"" Dec 03 00:38:17 crc kubenswrapper[4912]: I1203 00:38:17.376981 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" event={"ID":"25cacc35-89f1-4f0e-a621-ee002ce1f7e0","Type":"ContainerDied","Data":"e81015c0f5bbb5df99f28f6b91bbb221ffe9896190b645e0c8ffa2522432f620"} Dec 03 00:38:17 crc kubenswrapper[4912]: I1203 00:38:17.377062 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e81015c0f5bbb5df99f28f6b91bbb221ffe9896190b645e0c8ffa2522432f620" Dec 03 00:38:17 crc kubenswrapper[4912]: I1203 00:38:17.377079 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.912907 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn"] Dec 03 00:38:25 crc kubenswrapper[4912]: E1203 00:38:25.915876 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerName="util" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.916002 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerName="util" Dec 03 00:38:25 crc kubenswrapper[4912]: E1203 00:38:25.916089 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerName="extract" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.916162 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerName="extract" Dec 03 00:38:25 crc kubenswrapper[4912]: E1203 00:38:25.916245 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26654d73-149b-418b-8c5e-00653837545b" containerName="util" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.916401 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="26654d73-149b-418b-8c5e-00653837545b" containerName="util" Dec 03 00:38:25 crc kubenswrapper[4912]: E1203 00:38:25.916521 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26654d73-149b-418b-8c5e-00653837545b" containerName="extract" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.916600 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="26654d73-149b-418b-8c5e-00653837545b" containerName="extract" Dec 03 00:38:25 crc kubenswrapper[4912]: E1203 00:38:25.916678 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerName="pull" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.916749 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerName="pull" Dec 03 00:38:25 crc kubenswrapper[4912]: E1203 00:38:25.916847 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26654d73-149b-418b-8c5e-00653837545b" containerName="pull" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.916927 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="26654d73-149b-418b-8c5e-00653837545b" containerName="pull" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.917134 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="26654d73-149b-418b-8c5e-00653837545b" containerName="extract" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.917224 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="25cacc35-89f1-4f0e-a621-ee002ce1f7e0" containerName="extract" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.918329 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.925574 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.925734 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.925799 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.926661 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.926691 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.940185 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn"] Dec 03 00:38:25 crc kubenswrapper[4912]: I1203 00:38:25.940897 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-fxhg7" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.036182 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/32b6d349-0f43-45f8-93d6-c5f9f66273b2-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.036265 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32b6d349-0f43-45f8-93d6-c5f9f66273b2-webhook-cert\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.036354 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32b6d349-0f43-45f8-93d6-c5f9f66273b2-apiservice-cert\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.036514 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7cmk\" (UniqueName: \"kubernetes.io/projected/32b6d349-0f43-45f8-93d6-c5f9f66273b2-kube-api-access-m7cmk\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.036545 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/32b6d349-0f43-45f8-93d6-c5f9f66273b2-manager-config\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.137456 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32b6d349-0f43-45f8-93d6-c5f9f66273b2-apiservice-cert\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.137513 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32b6d349-0f43-45f8-93d6-c5f9f66273b2-webhook-cert\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.137575 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7cmk\" (UniqueName: \"kubernetes.io/projected/32b6d349-0f43-45f8-93d6-c5f9f66273b2-kube-api-access-m7cmk\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.137624 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/32b6d349-0f43-45f8-93d6-c5f9f66273b2-manager-config\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.137680 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/32b6d349-0f43-45f8-93d6-c5f9f66273b2-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.139350 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/32b6d349-0f43-45f8-93d6-c5f9f66273b2-manager-config\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.147768 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32b6d349-0f43-45f8-93d6-c5f9f66273b2-apiservice-cert\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.239894 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32b6d349-0f43-45f8-93d6-c5f9f66273b2-webhook-cert\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.240695 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/32b6d349-0f43-45f8-93d6-c5f9f66273b2-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.243674 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7cmk\" (UniqueName: \"kubernetes.io/projected/32b6d349-0f43-45f8-93d6-c5f9f66273b2-kube-api-access-m7cmk\") pod \"loki-operator-controller-manager-669f4ccbc6-wkghn\" (UID: \"32b6d349-0f43-45f8-93d6-c5f9f66273b2\") " pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.540952 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:26 crc kubenswrapper[4912]: I1203 00:38:26.837156 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn"] Dec 03 00:38:27 crc kubenswrapper[4912]: I1203 00:38:27.438421 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" event={"ID":"32b6d349-0f43-45f8-93d6-c5f9f66273b2","Type":"ContainerStarted","Data":"7d5ffdfebf234451ada287c7f1ef105a606a654422fb824dd9c5e47a62983ba7"} Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.285558 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-kxsph"] Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.286580 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.289129 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.289313 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.289456 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-wjljf" Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.305240 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-kxsph"] Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.394005 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnkc7\" (UniqueName: \"kubernetes.io/projected/fe13031d-b775-459e-ad64-7445ca8cbe63-kube-api-access-cnkc7\") pod \"cluster-logging-operator-ff9846bd-kxsph\" (UID: \"fe13031d-b775-459e-ad64-7445ca8cbe63\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.496882 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnkc7\" (UniqueName: \"kubernetes.io/projected/fe13031d-b775-459e-ad64-7445ca8cbe63-kube-api-access-cnkc7\") pod \"cluster-logging-operator-ff9846bd-kxsph\" (UID: \"fe13031d-b775-459e-ad64-7445ca8cbe63\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.526157 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnkc7\" (UniqueName: \"kubernetes.io/projected/fe13031d-b775-459e-ad64-7445ca8cbe63-kube-api-access-cnkc7\") pod \"cluster-logging-operator-ff9846bd-kxsph\" (UID: \"fe13031d-b775-459e-ad64-7445ca8cbe63\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" Dec 03 00:38:30 crc kubenswrapper[4912]: I1203 00:38:30.603567 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" Dec 03 00:38:31 crc kubenswrapper[4912]: I1203 00:38:31.087006 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-kxsph"] Dec 03 00:38:31 crc kubenswrapper[4912]: I1203 00:38:31.471367 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" event={"ID":"fe13031d-b775-459e-ad64-7445ca8cbe63","Type":"ContainerStarted","Data":"784a150191f202972dc21f8b4d5cea6748ae3d0113f2aa15eee078c23b981ee2"} Dec 03 00:38:37 crc kubenswrapper[4912]: I1203 00:38:37.549526 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" event={"ID":"32b6d349-0f43-45f8-93d6-c5f9f66273b2","Type":"ContainerStarted","Data":"2a6ac1e2fa8204fb791fc8173640fa11d0959affbedb0e7a0a506483c9ba99ba"} Dec 03 00:38:49 crc kubenswrapper[4912]: E1203 00:38:49.600473 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f15ebe396f96093861d528a3307a3e38ac2d4dff594f793c1e56011a3a909175" Dec 03 00:38:49 crc kubenswrapper[4912]: E1203 00:38:49.601642 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cluster-logging-operator,Image:registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f15ebe396f96093861d528a3307a3e38ac2d4dff594f793c1e56011a3a909175,Command:[cluster-logging-operator],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:cluster-logging-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_VECTOR,Value:registry.redhat.io/openshift-logging/vector-rhel9@sha256:438aa27c0408214bc64d01b20f233d698fa48344aa2dd878bc090232f227c17c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_LOG_FILE_METRIC_EXPORTER,Value:registry.redhat.io/openshift-logging/log-file-metric-exporter-rhel9@sha256:d4512c0a403fa3cddc646e566f5b6c77ac17c0558a08d8f99ffcdafaba9fba3a,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-logging.v6.2.6,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cnkc7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000690000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cluster-logging-operator-ff9846bd-kxsph_openshift-logging(fe13031d-b775-459e-ad64-7445ca8cbe63): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:38:49 crc kubenswrapper[4912]: E1203 00:38:49.602836 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-logging-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" podUID="fe13031d-b775-459e-ad64-7445ca8cbe63" Dec 03 00:38:49 crc kubenswrapper[4912]: E1203 00:38:49.981689 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-logging-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift-logging/cluster-logging-rhel9-operator@sha256:f15ebe396f96093861d528a3307a3e38ac2d4dff594f793c1e56011a3a909175\\\"\"" pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" podUID="fe13031d-b775-459e-ad64-7445ca8cbe63" Dec 03 00:38:57 crc kubenswrapper[4912]: E1203 00:38:57.836352 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24" Dec 03 00:38:57 crc kubenswrapper[4912]: E1203 00:38:57.837097 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --tls-cert-file=/var/run/secrets/serving-cert/tls.crt --tls-private-key-file=/var/run/secrets/serving-cert/tls.key --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256,TLS_RSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_256_GCM_SHA384,TLS_RSA_WITH_AES_128_CBC_SHA256 --tls-min-version=VersionTLS12 --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:loki-operator.v6.2.6,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:loki-operator-metrics-cert,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m7cmk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000700000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod loki-operator-controller-manager-669f4ccbc6-wkghn_openshift-operators-redhat(32b6d349-0f43-45f8-93d6-c5f9f66273b2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 00:38:57 crc kubenswrapper[4912]: E1203 00:38:57.838311 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" podUID="32b6d349-0f43-45f8-93d6-c5f9f66273b2" Dec 03 00:38:58 crc kubenswrapper[4912]: I1203 00:38:58.367320 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:58 crc kubenswrapper[4912]: I1203 00:38:58.369388 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" Dec 03 00:38:58 crc kubenswrapper[4912]: E1203 00:38:58.372883 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24\\\"\"" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" podUID="32b6d349-0f43-45f8-93d6-c5f9f66273b2" Dec 03 00:38:59 crc kubenswrapper[4912]: E1203 00:38:59.374851 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24\\\"\"" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" podUID="32b6d349-0f43-45f8-93d6-c5f9f66273b2" Dec 03 00:39:06 crc kubenswrapper[4912]: I1203 00:39:06.426222 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" event={"ID":"fe13031d-b775-459e-ad64-7445ca8cbe63","Type":"ContainerStarted","Data":"4ee1915a75ee0ff001f440a0a94d2201dd76cdf68e91e1ad3c01700d96076936"} Dec 03 00:39:16 crc kubenswrapper[4912]: I1203 00:39:16.489504 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" event={"ID":"32b6d349-0f43-45f8-93d6-c5f9f66273b2","Type":"ContainerStarted","Data":"c821c3ca8f8f0cfe802297dd5dbc9ecfb8bf3c820a3e7731125e1f56f1d2b58f"} Dec 03 00:39:16 crc kubenswrapper[4912]: I1203 00:39:16.512488 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-kxsph" podStartSLOduration=11.959638688 podStartE2EDuration="46.512473071s" podCreationTimestamp="2025-12-03 00:38:30 +0000 UTC" firstStartedPulling="2025-12-03 00:38:31.11834719 +0000 UTC m=+896.760367750" lastFinishedPulling="2025-12-03 00:39:05.671181573 +0000 UTC m=+931.313202133" observedRunningTime="2025-12-03 00:39:06.4485676 +0000 UTC m=+932.090588170" watchObservedRunningTime="2025-12-03 00:39:16.512473071 +0000 UTC m=+942.154493641" Dec 03 00:39:16 crc kubenswrapper[4912]: I1203 00:39:16.514318 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-669f4ccbc6-wkghn" podStartSLOduration=2.410650082 podStartE2EDuration="51.51431135s" podCreationTimestamp="2025-12-03 00:38:25 +0000 UTC" firstStartedPulling="2025-12-03 00:38:26.849848153 +0000 UTC m=+892.491868713" lastFinishedPulling="2025-12-03 00:39:15.953509421 +0000 UTC m=+941.595529981" observedRunningTime="2025-12-03 00:39:16.50976395 +0000 UTC m=+942.151784520" watchObservedRunningTime="2025-12-03 00:39:16.51431135 +0000 UTC m=+942.156331910" Dec 03 00:39:18 crc kubenswrapper[4912]: I1203 00:39:18.077266 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:39:18 crc kubenswrapper[4912]: I1203 00:39:18.077606 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.027636 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.028911 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.031714 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.031937 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.036598 4912 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-tdpbh" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.045284 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.171893 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-268effde-6056-4345-b211-76bf1c65a443\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-268effde-6056-4345-b211-76bf1c65a443\") pod \"minio\" (UID: \"77a7ed74-76f2-4541-8e49-7c2ebc3a11ef\") " pod="minio-dev/minio" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.171953 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj6vh\" (UniqueName: \"kubernetes.io/projected/77a7ed74-76f2-4541-8e49-7c2ebc3a11ef-kube-api-access-pj6vh\") pod \"minio\" (UID: \"77a7ed74-76f2-4541-8e49-7c2ebc3a11ef\") " pod="minio-dev/minio" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.272912 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-268effde-6056-4345-b211-76bf1c65a443\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-268effde-6056-4345-b211-76bf1c65a443\") pod \"minio\" (UID: \"77a7ed74-76f2-4541-8e49-7c2ebc3a11ef\") " pod="minio-dev/minio" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.272972 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj6vh\" (UniqueName: \"kubernetes.io/projected/77a7ed74-76f2-4541-8e49-7c2ebc3a11ef-kube-api-access-pj6vh\") pod \"minio\" (UID: \"77a7ed74-76f2-4541-8e49-7c2ebc3a11ef\") " pod="minio-dev/minio" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.277105 4912 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.277163 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-268effde-6056-4345-b211-76bf1c65a443\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-268effde-6056-4345-b211-76bf1c65a443\") pod \"minio\" (UID: \"77a7ed74-76f2-4541-8e49-7c2ebc3a11ef\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/83474a59534a95e5e4f99588e70d967304518d9827095aeffd115c909c0995cd/globalmount\"" pod="minio-dev/minio" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.295179 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj6vh\" (UniqueName: \"kubernetes.io/projected/77a7ed74-76f2-4541-8e49-7c2ebc3a11ef-kube-api-access-pj6vh\") pod \"minio\" (UID: \"77a7ed74-76f2-4541-8e49-7c2ebc3a11ef\") " pod="minio-dev/minio" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.305118 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-268effde-6056-4345-b211-76bf1c65a443\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-268effde-6056-4345-b211-76bf1c65a443\") pod \"minio\" (UID: \"77a7ed74-76f2-4541-8e49-7c2ebc3a11ef\") " pod="minio-dev/minio" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.353724 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 03 00:39:20 crc kubenswrapper[4912]: I1203 00:39:20.665322 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 03 00:39:21 crc kubenswrapper[4912]: I1203 00:39:21.521104 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"77a7ed74-76f2-4541-8e49-7c2ebc3a11ef","Type":"ContainerStarted","Data":"79072ec8d2cea0768eb1d83b71e4739cedc065a044ad881db4301382e77ab36e"} Dec 03 00:39:27 crc kubenswrapper[4912]: I1203 00:39:27.570160 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"77a7ed74-76f2-4541-8e49-7c2ebc3a11ef","Type":"ContainerStarted","Data":"909a8608fa1c1492aa3c64ce0e08d177fd81386df55beb6dabed4f12a08b9cef"} Dec 03 00:39:27 crc kubenswrapper[4912]: I1203 00:39:27.587003 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.193237623 podStartE2EDuration="10.586986949s" podCreationTimestamp="2025-12-03 00:39:17 +0000 UTC" firstStartedPulling="2025-12-03 00:39:20.680753383 +0000 UTC m=+946.322773943" lastFinishedPulling="2025-12-03 00:39:27.074502709 +0000 UTC m=+952.716523269" observedRunningTime="2025-12-03 00:39:27.584979305 +0000 UTC m=+953.226999865" watchObservedRunningTime="2025-12-03 00:39:27.586986949 +0000 UTC m=+953.229007509" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.322585 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.323648 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.328147 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.328388 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.328528 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-8ftdw" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.328636 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.328879 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.341590 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.501494 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aa94604-5389-4ef5-8e97-63167d5fb009-config\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.501573 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/3aa94604-5389-4ef5-8e97-63167d5fb009-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.501655 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7hvl\" (UniqueName: \"kubernetes.io/projected/3aa94604-5389-4ef5-8e97-63167d5fb009-kube-api-access-c7hvl\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.501695 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/3aa94604-5389-4ef5-8e97-63167d5fb009-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.501715 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa94604-5389-4ef5-8e97-63167d5fb009-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.506929 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-vxq5n"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.507893 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.510340 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.510490 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.512794 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.535273 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-vxq5n"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.602873 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.602936 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/3aa94604-5389-4ef5-8e97-63167d5fb009-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.602957 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa94604-5389-4ef5-8e97-63167d5fb009-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.602980 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxq7z\" (UniqueName: \"kubernetes.io/projected/3828ecda-60a7-4376-a5ec-f2730b95945d-kube-api-access-bxq7z\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.603106 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aa94604-5389-4ef5-8e97-63167d5fb009-config\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.603200 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.603248 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/3aa94604-5389-4ef5-8e97-63167d5fb009-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.603382 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.603414 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3828ecda-60a7-4376-a5ec-f2730b95945d-config\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.603528 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.603566 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7hvl\" (UniqueName: \"kubernetes.io/projected/3aa94604-5389-4ef5-8e97-63167d5fb009-kube-api-access-c7hvl\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.604637 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa94604-5389-4ef5-8e97-63167d5fb009-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.605366 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3aa94604-5389-4ef5-8e97-63167d5fb009-config\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.612258 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/3aa94604-5389-4ef5-8e97-63167d5fb009-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.614226 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.614997 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.618533 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.618602 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.623296 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7hvl\" (UniqueName: \"kubernetes.io/projected/3aa94604-5389-4ef5-8e97-63167d5fb009-kube-api-access-c7hvl\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.636690 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.647055 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/3aa94604-5389-4ef5-8e97-63167d5fb009-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-cr4z6\" (UID: \"3aa94604-5389-4ef5-8e97-63167d5fb009\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.663858 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705528 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaba064-9cc2-44fe-8516-29b31630da75-config\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705577 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxq7z\" (UniqueName: \"kubernetes.io/projected/3828ecda-60a7-4376-a5ec-f2730b95945d-kube-api-access-bxq7z\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705626 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705715 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faaba064-9cc2-44fe-8516-29b31630da75-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705739 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/faaba064-9cc2-44fe-8516-29b31630da75-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705782 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705802 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3828ecda-60a7-4376-a5ec-f2730b95945d-config\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705867 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705894 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/faaba064-9cc2-44fe-8516-29b31630da75-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705938 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.705965 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m594p\" (UniqueName: \"kubernetes.io/projected/faaba064-9cc2-44fe-8516-29b31630da75-kube-api-access-m594p\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.708683 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.714643 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3828ecda-60a7-4376-a5ec-f2730b95945d-config\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.717755 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.719179 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.732520 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/3828ecda-60a7-4376-a5ec-f2730b95945d-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.738339 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxq7z\" (UniqueName: \"kubernetes.io/projected/3828ecda-60a7-4376-a5ec-f2730b95945d-kube-api-access-bxq7z\") pod \"logging-loki-querier-5895d59bb8-vxq5n\" (UID: \"3828ecda-60a7-4376-a5ec-f2730b95945d\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.752502 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-777c467b8d-zsrn8"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.753857 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.765112 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-777c467b8d-zsrn8"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.765843 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.765863 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.766243 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-7mxj8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.767058 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.767213 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.767349 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.769873 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-777c467b8d-tfg7h"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.772091 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.791294 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-777c467b8d-tfg7h"] Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.807297 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faaba064-9cc2-44fe-8516-29b31630da75-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.807370 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/faaba064-9cc2-44fe-8516-29b31630da75-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.807466 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/faaba064-9cc2-44fe-8516-29b31630da75-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.807517 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m594p\" (UniqueName: \"kubernetes.io/projected/faaba064-9cc2-44fe-8516-29b31630da75-kube-api-access-m594p\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.807562 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaba064-9cc2-44fe-8516-29b31630da75-config\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.809049 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/faaba064-9cc2-44fe-8516-29b31630da75-config\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.810324 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faaba064-9cc2-44fe-8516-29b31630da75-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.818605 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/faaba064-9cc2-44fe-8516-29b31630da75-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.831094 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.849489 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/faaba064-9cc2-44fe-8516-29b31630da75-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.863134 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m594p\" (UniqueName: \"kubernetes.io/projected/faaba064-9cc2-44fe-8516-29b31630da75-kube-api-access-m594p\") pod \"logging-loki-query-frontend-84558f7c9f-ljkc8\" (UID: \"faaba064-9cc2-44fe-8516-29b31630da75\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.942322 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/00d2fcad-0e8b-4daf-ae2d-084456cd7144-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.942532 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-lokistack-gateway\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.942624 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-rbac\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.942696 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-logging-loki-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.942793 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/00d2fcad-0e8b-4daf-ae2d-084456cd7144-tls-secret\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.942874 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.942943 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/dd4ddd07-754f-48f9-b082-97b948a8d041-tenants\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.943236 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-lokistack-gateway\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.943312 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/dd4ddd07-754f-48f9-b082-97b948a8d041-tls-secret\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.943411 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/dd4ddd07-754f-48f9-b082-97b948a8d041-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.943509 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-rbac\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.943578 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fc9z\" (UniqueName: \"kubernetes.io/projected/00d2fcad-0e8b-4daf-ae2d-084456cd7144-kube-api-access-2fc9z\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.943649 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-logging-loki-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.943768 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8ks5\" (UniqueName: \"kubernetes.io/projected/dd4ddd07-754f-48f9-b082-97b948a8d041-kube-api-access-m8ks5\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.944103 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/00d2fcad-0e8b-4daf-ae2d-084456cd7144-tenants\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:32 crc kubenswrapper[4912]: I1203 00:39:32.944198 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.045949 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-rbac\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046004 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fc9z\" (UniqueName: \"kubernetes.io/projected/00d2fcad-0e8b-4daf-ae2d-084456cd7144-kube-api-access-2fc9z\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046025 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-logging-loki-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046061 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8ks5\" (UniqueName: \"kubernetes.io/projected/dd4ddd07-754f-48f9-b082-97b948a8d041-kube-api-access-m8ks5\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046077 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/00d2fcad-0e8b-4daf-ae2d-084456cd7144-tenants\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046139 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046180 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/00d2fcad-0e8b-4daf-ae2d-084456cd7144-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046229 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-lokistack-gateway\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046252 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-rbac\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046272 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-logging-loki-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046333 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/00d2fcad-0e8b-4daf-ae2d-084456cd7144-tls-secret\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046379 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046404 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/dd4ddd07-754f-48f9-b082-97b948a8d041-tenants\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046455 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-lokistack-gateway\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046478 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/dd4ddd07-754f-48f9-b082-97b948a8d041-tls-secret\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.046526 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/dd4ddd07-754f-48f9-b082-97b948a8d041-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.048348 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-lokistack-gateway\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.050909 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-rbac\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.051035 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.051995 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-rbac\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.052522 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-logging-loki-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.052699 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dd4ddd07-754f-48f9-b082-97b948a8d041-logging-loki-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.056115 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.059191 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/dd4ddd07-754f-48f9-b082-97b948a8d041-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.064695 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/dd4ddd07-754f-48f9-b082-97b948a8d041-tenants\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.075278 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/dd4ddd07-754f-48f9-b082-97b948a8d041-tls-secret\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.075467 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/00d2fcad-0e8b-4daf-ae2d-084456cd7144-lokistack-gateway\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.078958 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/00d2fcad-0e8b-4daf-ae2d-084456cd7144-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.109068 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/00d2fcad-0e8b-4daf-ae2d-084456cd7144-tenants\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.112712 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.112992 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8ks5\" (UniqueName: \"kubernetes.io/projected/dd4ddd07-754f-48f9-b082-97b948a8d041-kube-api-access-m8ks5\") pod \"logging-loki-gateway-777c467b8d-zsrn8\" (UID: \"dd4ddd07-754f-48f9-b082-97b948a8d041\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.113110 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fc9z\" (UniqueName: \"kubernetes.io/projected/00d2fcad-0e8b-4daf-ae2d-084456cd7144-kube-api-access-2fc9z\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.113515 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/00d2fcad-0e8b-4daf-ae2d-084456cd7144-tls-secret\") pod \"logging-loki-gateway-777c467b8d-tfg7h\" (UID: \"00d2fcad-0e8b-4daf-ae2d-084456cd7144\") " pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.247891 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.264107 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.553659 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.554873 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.557198 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.557495 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.561520 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.563580 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.571574 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.571780 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.590191 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.631857 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6"] Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.716269 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.733507 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.734470 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.743699 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.743804 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.750890 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762503 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762548 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762570 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762592 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e70879-0df5-4529-969f-e88309e22cb0-config\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762608 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94vxj\" (UniqueName: \"kubernetes.io/projected/2ead2e0e-573f-4ef4-8122-34c91de2fabe-kube-api-access-94vxj\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762631 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ead2e0e-573f-4ef4-8122-34c91de2fabe-config\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762650 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762674 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762700 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762723 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762744 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-47294351-251a-4e5b-a015-8ae4142437f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47294351-251a-4e5b-a015-8ae4142437f9\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762769 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762789 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-554db5d7-5272-4a73-a9f7-e705f876226f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-554db5d7-5272-4a73-a9f7-e705f876226f\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762806 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrqb6\" (UniqueName: \"kubernetes.io/projected/33e70879-0df5-4529-969f-e88309e22cb0-kube-api-access-zrqb6\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.762827 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f2966dff-e519-47ba-90fb-c440f0718e2e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f2966dff-e519-47ba-90fb-c440f0718e2e\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866656 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ead2e0e-573f-4ef4-8122-34c91de2fabe-config\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866701 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866733 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866759 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866786 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866806 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-47294351-251a-4e5b-a015-8ae4142437f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47294351-251a-4e5b-a015-8ae4142437f9\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866829 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866850 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-554db5d7-5272-4a73-a9f7-e705f876226f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-554db5d7-5272-4a73-a9f7-e705f876226f\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866867 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrqb6\" (UniqueName: \"kubernetes.io/projected/33e70879-0df5-4529-969f-e88309e22cb0-kube-api-access-zrqb6\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866901 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f2966dff-e519-47ba-90fb-c440f0718e2e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f2966dff-e519-47ba-90fb-c440f0718e2e\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866931 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866954 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.866983 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.867004 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e70879-0df5-4529-969f-e88309e22cb0-config\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.867022 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94vxj\" (UniqueName: \"kubernetes.io/projected/2ead2e0e-573f-4ef4-8122-34c91de2fabe-kube-api-access-94vxj\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.868146 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ead2e0e-573f-4ef4-8122-34c91de2fabe-config\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.872666 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.873471 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.875072 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33e70879-0df5-4529-969f-e88309e22cb0-config\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.881825 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.886655 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.889420 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.891233 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.894328 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/33e70879-0df5-4529-969f-e88309e22cb0-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.925372 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/2ead2e0e-573f-4ef4-8122-34c91de2fabe-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.925534 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrqb6\" (UniqueName: \"kubernetes.io/projected/33e70879-0df5-4529-969f-e88309e22cb0-kube-api-access-zrqb6\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.926133 4912 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.926196 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-554db5d7-5272-4a73-a9f7-e705f876226f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-554db5d7-5272-4a73-a9f7-e705f876226f\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8d8f2004ae1a60123dbe0ca4a7a504ba566bb98c916f8f704ea4ba2f15b52224/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.929072 4912 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.929113 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f2966dff-e519-47ba-90fb-c440f0718e2e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f2966dff-e519-47ba-90fb-c440f0718e2e\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/15d015f65159daccd36d4728dbededfabb86b2b59735814eee47ebb3cd0d43d7/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.929280 4912 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.929313 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-47294351-251a-4e5b-a015-8ae4142437f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47294351-251a-4e5b-a015-8ae4142437f9\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8a9c6f829363982417d6bdd6d795c253e2bda86446978284cb10b98dba7a53ba/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.934108 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-vxq5n"] Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.934139 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94vxj\" (UniqueName: \"kubernetes.io/projected/2ead2e0e-573f-4ef4-8122-34c91de2fabe-kube-api-access-94vxj\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.977214 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d7c01cab-ae55-43e7-9205-0abd19a99ca7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d7c01cab-ae55-43e7-9205-0abd19a99ca7\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.977287 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-config\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.977312 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.977338 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndcq8\" (UniqueName: \"kubernetes.io/projected/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-kube-api-access-ndcq8\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.977361 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.977393 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:33 crc kubenswrapper[4912]: I1203 00:39:33.977419 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.012264 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f2966dff-e519-47ba-90fb-c440f0718e2e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f2966dff-e519-47ba-90fb-c440f0718e2e\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.035367 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-554db5d7-5272-4a73-a9f7-e705f876226f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-554db5d7-5272-4a73-a9f7-e705f876226f\") pod \"logging-loki-ingester-0\" (UID: \"33e70879-0df5-4529-969f-e88309e22cb0\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.055933 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-47294351-251a-4e5b-a015-8ae4142437f9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-47294351-251a-4e5b-a015-8ae4142437f9\") pod \"logging-loki-compactor-0\" (UID: \"2ead2e0e-573f-4ef4-8122-34c91de2fabe\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.078596 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.078651 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.078700 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d7c01cab-ae55-43e7-9205-0abd19a99ca7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d7c01cab-ae55-43e7-9205-0abd19a99ca7\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.078731 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-config\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.078772 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.078797 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndcq8\" (UniqueName: \"kubernetes.io/projected/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-kube-api-access-ndcq8\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.078815 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.080141 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-config\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.081194 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.081973 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.082793 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.086090 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.086353 4912 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.086379 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d7c01cab-ae55-43e7-9205-0abd19a99ca7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d7c01cab-ae55-43e7-9205-0abd19a99ca7\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9fc2595c2ebdd06c8e0830c4eb12fc4443c6409bea79fd83d8a65ae4a18fd729/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.098061 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndcq8\" (UniqueName: \"kubernetes.io/projected/dfc8d0a7-a087-4dfb-a460-e63aeb64828e-kube-api-access-ndcq8\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.121776 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d7c01cab-ae55-43e7-9205-0abd19a99ca7\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d7c01cab-ae55-43e7-9205-0abd19a99ca7\") pod \"logging-loki-index-gateway-0\" (UID: \"dfc8d0a7-a087-4dfb-a460-e63aeb64828e\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.175912 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.178174 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-777c467b8d-zsrn8"] Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.191821 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8"] Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.199477 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:34 crc kubenswrapper[4912]: W1203 00:39:34.202330 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaaba064_9cc2_44fe_8516_29b31630da75.slice/crio-bfd95f7d584f05da559882bb5269ac2266a8a2c4ae1241d7e5ad7885efe34b80 WatchSource:0}: Error finding container bfd95f7d584f05da559882bb5269ac2266a8a2c4ae1241d7e5ad7885efe34b80: Status 404 returned error can't find the container with id bfd95f7d584f05da559882bb5269ac2266a8a2c4ae1241d7e5ad7885efe34b80 Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.300897 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-777c467b8d-tfg7h"] Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.367522 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.442313 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 03 00:39:34 crc kubenswrapper[4912]: W1203 00:39:34.455700 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ead2e0e_573f_4ef4_8122_34c91de2fabe.slice/crio-4e47763fb37d05c537653e9c07fbaaf23d07d1eaa1a8baede9460ea227bdda9e WatchSource:0}: Error finding container 4e47763fb37d05c537653e9c07fbaaf23d07d1eaa1a8baede9460ea227bdda9e: Status 404 returned error can't find the container with id 4e47763fb37d05c537653e9c07fbaaf23d07d1eaa1a8baede9460ea227bdda9e Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.622845 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.626677 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.710532 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" event={"ID":"dd4ddd07-754f-48f9-b082-97b948a8d041","Type":"ContainerStarted","Data":"427b27a06f8a2bbaefc42f6f14241b35811959926dc4cbbbab3f09bfab6fda30"} Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.711811 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"dfc8d0a7-a087-4dfb-a460-e63aeb64828e","Type":"ContainerStarted","Data":"f494fa95003d5bf1e5dc677c26a8316f0ec5ca2df72f5100cc34ec37548083e9"} Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.713692 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"33e70879-0df5-4529-969f-e88309e22cb0","Type":"ContainerStarted","Data":"4b652beed93fceea529e9ba020ee5b76f9a98590286c38b79153f9066b5bd014"} Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.716083 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" event={"ID":"00d2fcad-0e8b-4daf-ae2d-084456cd7144","Type":"ContainerStarted","Data":"93850edd395679489822d97aff2919d944f53a0ad69d4936ee406e60a27224ad"} Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.718449 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" event={"ID":"3828ecda-60a7-4376-a5ec-f2730b95945d","Type":"ContainerStarted","Data":"6df16632147a9345a72e0b9bdf9b4bfd10932f1ffec2babb4ca51911af201fe1"} Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.726603 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" event={"ID":"faaba064-9cc2-44fe-8516-29b31630da75","Type":"ContainerStarted","Data":"bfd95f7d584f05da559882bb5269ac2266a8a2c4ae1241d7e5ad7885efe34b80"} Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.728664 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"2ead2e0e-573f-4ef4-8122-34c91de2fabe","Type":"ContainerStarted","Data":"4e47763fb37d05c537653e9c07fbaaf23d07d1eaa1a8baede9460ea227bdda9e"} Dec 03 00:39:34 crc kubenswrapper[4912]: I1203 00:39:34.731501 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" event={"ID":"3aa94604-5389-4ef5-8e97-63167d5fb009","Type":"ContainerStarted","Data":"28735963e90571fdad569ed675856ce7b96214150d3a0986d5c792455df1f258"} Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.014894 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7sz66"] Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.017186 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.041127 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7sz66"] Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.206952 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxxwz\" (UniqueName: \"kubernetes.io/projected/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-kube-api-access-hxxwz\") pod \"community-operators-7sz66\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.207346 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-utilities\") pod \"community-operators-7sz66\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.207448 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-catalog-content\") pod \"community-operators-7sz66\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.310952 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-utilities\") pod \"community-operators-7sz66\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.311041 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-catalog-content\") pod \"community-operators-7sz66\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.311189 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxxwz\" (UniqueName: \"kubernetes.io/projected/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-kube-api-access-hxxwz\") pod \"community-operators-7sz66\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.311594 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-catalog-content\") pod \"community-operators-7sz66\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.311747 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-utilities\") pod \"community-operators-7sz66\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.355253 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxxwz\" (UniqueName: \"kubernetes.io/projected/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-kube-api-access-hxxwz\") pod \"community-operators-7sz66\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:35 crc kubenswrapper[4912]: I1203 00:39:35.642904 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:36 crc kubenswrapper[4912]: I1203 00:39:36.015159 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7sz66"] Dec 03 00:39:36 crc kubenswrapper[4912]: I1203 00:39:36.802336 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7sz66" event={"ID":"123154a6-35bb-4f9c-80c3-db4ef0fcdc82","Type":"ContainerStarted","Data":"f639a1c767777053c42e804b86a1d628cb3def0bbe1c704568b51b6ff9e18bdb"} Dec 03 00:39:36 crc kubenswrapper[4912]: I1203 00:39:36.803256 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7sz66" event={"ID":"123154a6-35bb-4f9c-80c3-db4ef0fcdc82","Type":"ContainerStarted","Data":"83ef92dccace586c00eb227e836f7e4d16c5e4bda440d35f96e953ae1e390bac"} Dec 03 00:39:37 crc kubenswrapper[4912]: I1203 00:39:37.814655 4912 generic.go:334] "Generic (PLEG): container finished" podID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerID="f639a1c767777053c42e804b86a1d628cb3def0bbe1c704568b51b6ff9e18bdb" exitCode=0 Dec 03 00:39:37 crc kubenswrapper[4912]: I1203 00:39:37.814737 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7sz66" event={"ID":"123154a6-35bb-4f9c-80c3-db4ef0fcdc82","Type":"ContainerDied","Data":"f639a1c767777053c42e804b86a1d628cb3def0bbe1c704568b51b6ff9e18bdb"} Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.456934 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lbv9m"] Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.459349 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.466330 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-utilities\") pod \"certified-operators-lbv9m\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.466405 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-catalog-content\") pod \"certified-operators-lbv9m\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.466883 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5qks\" (UniqueName: \"kubernetes.io/projected/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-kube-api-access-b5qks\") pod \"certified-operators-lbv9m\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.473560 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lbv9m"] Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.568327 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-utilities\") pod \"certified-operators-lbv9m\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.568405 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-catalog-content\") pod \"certified-operators-lbv9m\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.568484 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5qks\" (UniqueName: \"kubernetes.io/projected/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-kube-api-access-b5qks\") pod \"certified-operators-lbv9m\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.568838 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-utilities\") pod \"certified-operators-lbv9m\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.569058 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-catalog-content\") pod \"certified-operators-lbv9m\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.590423 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5qks\" (UniqueName: \"kubernetes.io/projected/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-kube-api-access-b5qks\") pod \"certified-operators-lbv9m\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:42 crc kubenswrapper[4912]: I1203 00:39:42.785104 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.263947 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lbv9m"] Dec 03 00:39:43 crc kubenswrapper[4912]: W1203 00:39:43.297580 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7d77826_4066_4f54_bdb9_e1fbbafdfa4c.slice/crio-b66e3c1662b8272fe8d2e39ad9fbfb47d6cfc2b370f6dc5cc173cdf64fe9677f WatchSource:0}: Error finding container b66e3c1662b8272fe8d2e39ad9fbfb47d6cfc2b370f6dc5cc173cdf64fe9677f: Status 404 returned error can't find the container with id b66e3c1662b8272fe8d2e39ad9fbfb47d6cfc2b370f6dc5cc173cdf64fe9677f Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.895827 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" event={"ID":"3828ecda-60a7-4376-a5ec-f2730b95945d","Type":"ContainerStarted","Data":"b1a8290725727a05bf036fe89bca41fe5a3779180a7f254d85d2b3017938db93"} Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.896417 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.899418 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"33e70879-0df5-4529-969f-e88309e22cb0","Type":"ContainerStarted","Data":"9107e9ac5064872ea69ccac447045c6311c87585ffeba7517ffc3a56ae63f84f"} Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.936524 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.955456 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbv9m" event={"ID":"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c","Type":"ContainerStarted","Data":"b66e3c1662b8272fe8d2e39ad9fbfb47d6cfc2b370f6dc5cc173cdf64fe9677f"} Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.959294 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" podStartSLOduration=3.032032015 podStartE2EDuration="11.959280895s" podCreationTimestamp="2025-12-03 00:39:32 +0000 UTC" firstStartedPulling="2025-12-03 00:39:33.955758542 +0000 UTC m=+959.597779102" lastFinishedPulling="2025-12-03 00:39:42.883007422 +0000 UTC m=+968.525027982" observedRunningTime="2025-12-03 00:39:43.957253909 +0000 UTC m=+969.599274469" watchObservedRunningTime="2025-12-03 00:39:43.959280895 +0000 UTC m=+969.601301455" Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.978632 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" event={"ID":"faaba064-9cc2-44fe-8516-29b31630da75","Type":"ContainerStarted","Data":"7aa568a8a24c5019cc3dd0bf556d176acef523042843a43c6e4cb35f3a3c00af"} Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.979648 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:39:43 crc kubenswrapper[4912]: I1203 00:39:43.994521 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7sz66" event={"ID":"123154a6-35bb-4f9c-80c3-db4ef0fcdc82","Type":"ContainerStarted","Data":"13d25b0e0c99dfcb644553dfdacd05afc39bbf169ae8262d91b2872962cd8bee"} Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.008048 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.011531 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" event={"ID":"3aa94604-5389-4ef5-8e97-63167d5fb009","Type":"ContainerStarted","Data":"b09dccf56ebe55dcc54883eae3dcdbc030c08d20fe7d107ac38131e690dd9fe8"} Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.012019 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.014414 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" event={"ID":"dd4ddd07-754f-48f9-b082-97b948a8d041","Type":"ContainerStarted","Data":"b5718b266a6ffd120846c65968a4fe30f312273b738f067fd44d8f9f76c270f5"} Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.017118 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.024242 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=4.615857325 podStartE2EDuration="12.024213393s" podCreationTimestamp="2025-12-03 00:39:32 +0000 UTC" firstStartedPulling="2025-12-03 00:39:34.647706364 +0000 UTC m=+960.289726924" lastFinishedPulling="2025-12-03 00:39:42.056062432 +0000 UTC m=+967.698082992" observedRunningTime="2025-12-03 00:39:44.019802174 +0000 UTC m=+969.661822744" watchObservedRunningTime="2025-12-03 00:39:44.024213393 +0000 UTC m=+969.666233953" Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.049176 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.621868661 podStartE2EDuration="12.049145128s" podCreationTimestamp="2025-12-03 00:39:32 +0000 UTC" firstStartedPulling="2025-12-03 00:39:34.461057939 +0000 UTC m=+960.103078499" lastFinishedPulling="2025-12-03 00:39:42.888334406 +0000 UTC m=+968.530354966" observedRunningTime="2025-12-03 00:39:44.04328086 +0000 UTC m=+969.685301420" watchObservedRunningTime="2025-12-03 00:39:44.049145128 +0000 UTC m=+969.691165688" Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.068635 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" podStartSLOduration=2.888472036 podStartE2EDuration="12.068613116s" podCreationTimestamp="2025-12-03 00:39:32 +0000 UTC" firstStartedPulling="2025-12-03 00:39:33.706834769 +0000 UTC m=+959.348855329" lastFinishedPulling="2025-12-03 00:39:42.886975849 +0000 UTC m=+968.528996409" observedRunningTime="2025-12-03 00:39:44.067940697 +0000 UTC m=+969.709961257" watchObservedRunningTime="2025-12-03 00:39:44.068613116 +0000 UTC m=+969.710633676" Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.093960 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" podStartSLOduration=3.420833836 podStartE2EDuration="12.093944472s" podCreationTimestamp="2025-12-03 00:39:32 +0000 UTC" firstStartedPulling="2025-12-03 00:39:34.209233638 +0000 UTC m=+959.851254198" lastFinishedPulling="2025-12-03 00:39:42.882344264 +0000 UTC m=+968.524364834" observedRunningTime="2025-12-03 00:39:44.091860235 +0000 UTC m=+969.733880805" watchObservedRunningTime="2025-12-03 00:39:44.093944472 +0000 UTC m=+969.735965032" Dec 03 00:39:44 crc kubenswrapper[4912]: I1203 00:39:44.135162 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.887383984 podStartE2EDuration="12.135146478s" podCreationTimestamp="2025-12-03 00:39:32 +0000 UTC" firstStartedPulling="2025-12-03 00:39:34.632615006 +0000 UTC m=+960.274635566" lastFinishedPulling="2025-12-03 00:39:42.8803775 +0000 UTC m=+968.522398060" observedRunningTime="2025-12-03 00:39:44.131614112 +0000 UTC m=+969.773634672" watchObservedRunningTime="2025-12-03 00:39:44.135146478 +0000 UTC m=+969.777167038" Dec 03 00:39:45 crc kubenswrapper[4912]: I1203 00:39:45.025581 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"2ead2e0e-573f-4ef4-8122-34c91de2fabe","Type":"ContainerStarted","Data":"8784815a905505ee4ccd5e3376385fe03c527a84bc2a5b6000b47e966719b040"} Dec 03 00:39:45 crc kubenswrapper[4912]: I1203 00:39:45.027653 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"dfc8d0a7-a087-4dfb-a460-e63aeb64828e","Type":"ContainerStarted","Data":"57aedec0efc81d58cc354db5df52971d0d969dab7b266a859c27800cb25fdb8a"} Dec 03 00:39:45 crc kubenswrapper[4912]: I1203 00:39:45.029062 4912 generic.go:334] "Generic (PLEG): container finished" podID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerID="1723885e60187b448ad9bfc53b0dde1b1f0d18932c0edd234db8f48ec1e90bab" exitCode=0 Dec 03 00:39:45 crc kubenswrapper[4912]: I1203 00:39:45.029187 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbv9m" event={"ID":"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c","Type":"ContainerDied","Data":"1723885e60187b448ad9bfc53b0dde1b1f0d18932c0edd234db8f48ec1e90bab"} Dec 03 00:39:45 crc kubenswrapper[4912]: I1203 00:39:45.030723 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" event={"ID":"00d2fcad-0e8b-4daf-ae2d-084456cd7144","Type":"ContainerStarted","Data":"06bc3ae058339144366a1287f193e9cc655a52ca040eb92a3ac0f5af081ccca5"} Dec 03 00:39:45 crc kubenswrapper[4912]: I1203 00:39:45.032789 4912 generic.go:334] "Generic (PLEG): container finished" podID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerID="13d25b0e0c99dfcb644553dfdacd05afc39bbf169ae8262d91b2872962cd8bee" exitCode=0 Dec 03 00:39:45 crc kubenswrapper[4912]: I1203 00:39:45.032859 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7sz66" event={"ID":"123154a6-35bb-4f9c-80c3-db4ef0fcdc82","Type":"ContainerDied","Data":"13d25b0e0c99dfcb644553dfdacd05afc39bbf169ae8262d91b2872962cd8bee"} Dec 03 00:39:46 crc kubenswrapper[4912]: I1203 00:39:46.043746 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbv9m" event={"ID":"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c","Type":"ContainerStarted","Data":"0de3bc9625cbd655479d8e93a4bcd8acad44b535c43ed94a71bcf124fe5d773f"} Dec 03 00:39:46 crc kubenswrapper[4912]: E1203 00:39:46.350923 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7d77826_4066_4f54_bdb9_e1fbbafdfa4c.slice/crio-conmon-0de3bc9625cbd655479d8e93a4bcd8acad44b535c43ed94a71bcf124fe5d773f.scope\": RecentStats: unable to find data in memory cache]" Dec 03 00:39:47 crc kubenswrapper[4912]: I1203 00:39:47.055634 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7sz66" event={"ID":"123154a6-35bb-4f9c-80c3-db4ef0fcdc82","Type":"ContainerStarted","Data":"249b5f11828b3049832ef6a8ef26424ac9460ad10888cd096e0450c16715cdb6"} Dec 03 00:39:47 crc kubenswrapper[4912]: I1203 00:39:47.078027 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7sz66" podStartSLOduration=6.859552363 podStartE2EDuration="13.07800304s" podCreationTimestamp="2025-12-03 00:39:34 +0000 UTC" firstStartedPulling="2025-12-03 00:39:39.688881693 +0000 UTC m=+965.330902273" lastFinishedPulling="2025-12-03 00:39:45.90733239 +0000 UTC m=+971.549352950" observedRunningTime="2025-12-03 00:39:47.07799357 +0000 UTC m=+972.720014130" watchObservedRunningTime="2025-12-03 00:39:47.07800304 +0000 UTC m=+972.720023600" Dec 03 00:39:47 crc kubenswrapper[4912]: I1203 00:39:47.078550 4912 generic.go:334] "Generic (PLEG): container finished" podID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerID="0de3bc9625cbd655479d8e93a4bcd8acad44b535c43ed94a71bcf124fe5d773f" exitCode=0 Dec 03 00:39:47 crc kubenswrapper[4912]: I1203 00:39:47.078629 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbv9m" event={"ID":"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c","Type":"ContainerDied","Data":"0de3bc9625cbd655479d8e93a4bcd8acad44b535c43ed94a71bcf124fe5d773f"} Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.077784 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.078221 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.088680 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbv9m" event={"ID":"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c","Type":"ContainerStarted","Data":"91e8e0b6c6ed4d3a700474a6ac09ebf880e314ae33c9f31a393a47e1eaf5b870"} Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.090572 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" event={"ID":"00d2fcad-0e8b-4daf-ae2d-084456cd7144","Type":"ContainerStarted","Data":"0922fa7c42407e0a0f50640fb5686228ec72948d20067f5eb89961cef2e4e760"} Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.090846 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.092415 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" event={"ID":"dd4ddd07-754f-48f9-b082-97b948a8d041","Type":"ContainerStarted","Data":"7dad83812d0d1a45c5fd06a2e539fad37e8660a1d6596ca8c93d661d4c798623"} Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.105921 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.114361 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lbv9m" podStartSLOduration=2.5190807079999997 podStartE2EDuration="6.114326231s" podCreationTimestamp="2025-12-03 00:39:42 +0000 UTC" firstStartedPulling="2025-12-03 00:39:43.96169527 +0000 UTC m=+969.603715830" lastFinishedPulling="2025-12-03 00:39:47.556940793 +0000 UTC m=+973.198961353" observedRunningTime="2025-12-03 00:39:48.108797251 +0000 UTC m=+973.750817821" watchObservedRunningTime="2025-12-03 00:39:48.114326231 +0000 UTC m=+973.756346781" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.132543 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" podStartSLOduration=3.256385372 podStartE2EDuration="16.132524053s" podCreationTimestamp="2025-12-03 00:39:32 +0000 UTC" firstStartedPulling="2025-12-03 00:39:34.317677186 +0000 UTC m=+959.959697746" lastFinishedPulling="2025-12-03 00:39:47.193815867 +0000 UTC m=+972.835836427" observedRunningTime="2025-12-03 00:39:48.12797248 +0000 UTC m=+973.769993060" watchObservedRunningTime="2025-12-03 00:39:48.132524053 +0000 UTC m=+973.774544613" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.160683 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" podStartSLOduration=3.15994931 podStartE2EDuration="16.160667856s" podCreationTimestamp="2025-12-03 00:39:32 +0000 UTC" firstStartedPulling="2025-12-03 00:39:34.188696742 +0000 UTC m=+959.830717302" lastFinishedPulling="2025-12-03 00:39:47.189415288 +0000 UTC m=+972.831435848" observedRunningTime="2025-12-03 00:39:48.158028885 +0000 UTC m=+973.800049465" watchObservedRunningTime="2025-12-03 00:39:48.160667856 +0000 UTC m=+973.802688406" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.248858 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.248931 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.264812 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.357929 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.404498 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-777c467b8d-tfg7h" Dec 03 00:39:48 crc kubenswrapper[4912]: I1203 00:39:48.405729 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-777c467b8d-zsrn8" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.444362 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5mmg5"] Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.447055 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.456988 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mmg5"] Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.626755 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-utilities\") pod \"redhat-marketplace-5mmg5\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.627586 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrkp8\" (UniqueName: \"kubernetes.io/projected/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-kube-api-access-mrkp8\") pod \"redhat-marketplace-5mmg5\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.628291 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-catalog-content\") pod \"redhat-marketplace-5mmg5\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.729819 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-utilities\") pod \"redhat-marketplace-5mmg5\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.729910 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrkp8\" (UniqueName: \"kubernetes.io/projected/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-kube-api-access-mrkp8\") pod \"redhat-marketplace-5mmg5\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.729952 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-catalog-content\") pod \"redhat-marketplace-5mmg5\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.730877 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-utilities\") pod \"redhat-marketplace-5mmg5\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.731520 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-catalog-content\") pod \"redhat-marketplace-5mmg5\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.754177 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrkp8\" (UniqueName: \"kubernetes.io/projected/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-kube-api-access-mrkp8\") pod \"redhat-marketplace-5mmg5\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:51 crc kubenswrapper[4912]: I1203 00:39:51.780864 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:39:52 crc kubenswrapper[4912]: I1203 00:39:52.048870 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mmg5"] Dec 03 00:39:52 crc kubenswrapper[4912]: I1203 00:39:52.123843 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mmg5" event={"ID":"f4577ec2-7bda-492f-8b34-cfbce20d5f5a","Type":"ContainerStarted","Data":"e7373bdb81f9b50ba8bbb0c9c25a406699b3a873853c84bc21738d29b87fdd5f"} Dec 03 00:39:52 crc kubenswrapper[4912]: I1203 00:39:52.785586 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:52 crc kubenswrapper[4912]: I1203 00:39:52.785652 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:52 crc kubenswrapper[4912]: I1203 00:39:52.831359 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:53 crc kubenswrapper[4912]: I1203 00:39:53.188685 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:54 crc kubenswrapper[4912]: I1203 00:39:54.139149 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerID="3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e" exitCode=0 Dec 03 00:39:54 crc kubenswrapper[4912]: I1203 00:39:54.141276 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mmg5" event={"ID":"f4577ec2-7bda-492f-8b34-cfbce20d5f5a","Type":"ContainerDied","Data":"3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e"} Dec 03 00:39:55 crc kubenswrapper[4912]: I1203 00:39:55.155865 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mmg5" event={"ID":"f4577ec2-7bda-492f-8b34-cfbce20d5f5a","Type":"ContainerStarted","Data":"cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461"} Dec 03 00:39:55 crc kubenswrapper[4912]: I1203 00:39:55.235981 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lbv9m"] Dec 03 00:39:55 crc kubenswrapper[4912]: I1203 00:39:55.236230 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lbv9m" podUID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerName="registry-server" containerID="cri-o://91e8e0b6c6ed4d3a700474a6ac09ebf880e314ae33c9f31a393a47e1eaf5b870" gracePeriod=2 Dec 03 00:39:55 crc kubenswrapper[4912]: I1203 00:39:55.643815 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:55 crc kubenswrapper[4912]: I1203 00:39:55.643903 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:55 crc kubenswrapper[4912]: I1203 00:39:55.699679 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.170459 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerID="cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461" exitCode=0 Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.170520 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mmg5" event={"ID":"f4577ec2-7bda-492f-8b34-cfbce20d5f5a","Type":"ContainerDied","Data":"cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461"} Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.174678 4912 generic.go:334] "Generic (PLEG): container finished" podID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerID="91e8e0b6c6ed4d3a700474a6ac09ebf880e314ae33c9f31a393a47e1eaf5b870" exitCode=0 Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.175182 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbv9m" event={"ID":"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c","Type":"ContainerDied","Data":"91e8e0b6c6ed4d3a700474a6ac09ebf880e314ae33c9f31a393a47e1eaf5b870"} Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.175215 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbv9m" event={"ID":"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c","Type":"ContainerDied","Data":"b66e3c1662b8272fe8d2e39ad9fbfb47d6cfc2b370f6dc5cc173cdf64fe9677f"} Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.175230 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b66e3c1662b8272fe8d2e39ad9fbfb47d6cfc2b370f6dc5cc173cdf64fe9677f" Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.179057 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.221988 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-catalog-content\") pod \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.222074 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-utilities\") pod \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.222195 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5qks\" (UniqueName: \"kubernetes.io/projected/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-kube-api-access-b5qks\") pod \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\" (UID: \"e7d77826-4066-4f54-bdb9-e1fbbafdfa4c\") " Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.223021 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-utilities" (OuterVolumeSpecName: "utilities") pod "e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" (UID: "e7d77826-4066-4f54-bdb9-e1fbbafdfa4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.229677 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.230289 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-kube-api-access-b5qks" (OuterVolumeSpecName: "kube-api-access-b5qks") pod "e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" (UID: "e7d77826-4066-4f54-bdb9-e1fbbafdfa4c"). InnerVolumeSpecName "kube-api-access-b5qks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.272560 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" (UID: "e7d77826-4066-4f54-bdb9-e1fbbafdfa4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.323902 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.323939 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:56 crc kubenswrapper[4912]: I1203 00:39:56.323980 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5qks\" (UniqueName: \"kubernetes.io/projected/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c-kube-api-access-b5qks\") on node \"crc\" DevicePath \"\"" Dec 03 00:39:57 crc kubenswrapper[4912]: I1203 00:39:57.184730 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mmg5" event={"ID":"f4577ec2-7bda-492f-8b34-cfbce20d5f5a","Type":"ContainerStarted","Data":"c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf"} Dec 03 00:39:57 crc kubenswrapper[4912]: I1203 00:39:57.184812 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbv9m" Dec 03 00:39:57 crc kubenswrapper[4912]: I1203 00:39:57.213370 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5mmg5" podStartSLOduration=3.774546675 podStartE2EDuration="6.213347785s" podCreationTimestamp="2025-12-03 00:39:51 +0000 UTC" firstStartedPulling="2025-12-03 00:39:54.142451884 +0000 UTC m=+979.784472444" lastFinishedPulling="2025-12-03 00:39:56.581252954 +0000 UTC m=+982.223273554" observedRunningTime="2025-12-03 00:39:57.21057236 +0000 UTC m=+982.852592930" watchObservedRunningTime="2025-12-03 00:39:57.213347785 +0000 UTC m=+982.855368345" Dec 03 00:39:57 crc kubenswrapper[4912]: I1203 00:39:57.240870 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lbv9m"] Dec 03 00:39:57 crc kubenswrapper[4912]: I1203 00:39:57.249296 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lbv9m"] Dec 03 00:39:58 crc kubenswrapper[4912]: I1203 00:39:58.579529 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" path="/var/lib/kubelet/pods/e7d77826-4066-4f54-bdb9-e1fbbafdfa4c/volumes" Dec 03 00:39:59 crc kubenswrapper[4912]: I1203 00:39:59.036990 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7sz66"] Dec 03 00:39:59 crc kubenswrapper[4912]: I1203 00:39:59.037740 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7sz66" podUID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerName="registry-server" containerID="cri-o://249b5f11828b3049832ef6a8ef26424ac9460ad10888cd096e0450c16715cdb6" gracePeriod=2 Dec 03 00:40:00 crc kubenswrapper[4912]: I1203 00:40:00.219384 4912 generic.go:334] "Generic (PLEG): container finished" podID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerID="249b5f11828b3049832ef6a8ef26424ac9460ad10888cd096e0450c16715cdb6" exitCode=0 Dec 03 00:40:00 crc kubenswrapper[4912]: I1203 00:40:00.219457 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7sz66" event={"ID":"123154a6-35bb-4f9c-80c3-db4ef0fcdc82","Type":"ContainerDied","Data":"249b5f11828b3049832ef6a8ef26424ac9460ad10888cd096e0450c16715cdb6"} Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.420365 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.507043 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-utilities\") pod \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.507098 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-catalog-content\") pod \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.507196 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxxwz\" (UniqueName: \"kubernetes.io/projected/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-kube-api-access-hxxwz\") pod \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\" (UID: \"123154a6-35bb-4f9c-80c3-db4ef0fcdc82\") " Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.508054 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-utilities" (OuterVolumeSpecName: "utilities") pod "123154a6-35bb-4f9c-80c3-db4ef0fcdc82" (UID: "123154a6-35bb-4f9c-80c3-db4ef0fcdc82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.518581 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-kube-api-access-hxxwz" (OuterVolumeSpecName: "kube-api-access-hxxwz") pod "123154a6-35bb-4f9c-80c3-db4ef0fcdc82" (UID: "123154a6-35bb-4f9c-80c3-db4ef0fcdc82"). InnerVolumeSpecName "kube-api-access-hxxwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.553177 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "123154a6-35bb-4f9c-80c3-db4ef0fcdc82" (UID: "123154a6-35bb-4f9c-80c3-db4ef0fcdc82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.608727 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.608771 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.608784 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxxwz\" (UniqueName: \"kubernetes.io/projected/123154a6-35bb-4f9c-80c3-db4ef0fcdc82-kube-api-access-hxxwz\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.781588 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.781645 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:40:01 crc kubenswrapper[4912]: I1203 00:40:01.820898 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.233422 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7sz66" event={"ID":"123154a6-35bb-4f9c-80c3-db4ef0fcdc82","Type":"ContainerDied","Data":"83ef92dccace586c00eb227e836f7e4d16c5e4bda440d35f96e953ae1e390bac"} Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.233523 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7sz66" Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.233911 4912 scope.go:117] "RemoveContainer" containerID="249b5f11828b3049832ef6a8ef26424ac9460ad10888cd096e0450c16715cdb6" Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.249774 4912 scope.go:117] "RemoveContainer" containerID="13d25b0e0c99dfcb644553dfdacd05afc39bbf169ae8262d91b2872962cd8bee" Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.272855 4912 scope.go:117] "RemoveContainer" containerID="f639a1c767777053c42e804b86a1d628cb3def0bbe1c704568b51b6ff9e18bdb" Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.282282 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7sz66"] Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.288532 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.292146 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7sz66"] Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.584657 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" path="/var/lib/kubelet/pods/123154a6-35bb-4f9c-80c3-db4ef0fcdc82/volumes" Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.671779 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-cr4z6" Dec 03 00:40:02 crc kubenswrapper[4912]: I1203 00:40:02.837558 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-vxq5n" Dec 03 00:40:03 crc kubenswrapper[4912]: I1203 00:40:03.119608 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-ljkc8" Dec 03 00:40:04 crc kubenswrapper[4912]: I1203 00:40:04.184566 4912 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 03 00:40:04 crc kubenswrapper[4912]: I1203 00:40:04.184623 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="33e70879-0df5-4529-969f-e88309e22cb0" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 00:40:04 crc kubenswrapper[4912]: I1203 00:40:04.205882 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 03 00:40:04 crc kubenswrapper[4912]: I1203 00:40:04.242264 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mmg5"] Dec 03 00:40:04 crc kubenswrapper[4912]: I1203 00:40:04.258756 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5mmg5" podUID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerName="registry-server" containerID="cri-o://c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf" gracePeriod=2 Dec 03 00:40:04 crc kubenswrapper[4912]: I1203 00:40:04.378106 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.203689 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.268296 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerID="c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf" exitCode=0 Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.268384 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mmg5" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.268347 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mmg5" event={"ID":"f4577ec2-7bda-492f-8b34-cfbce20d5f5a","Type":"ContainerDied","Data":"c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf"} Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.268550 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mmg5" event={"ID":"f4577ec2-7bda-492f-8b34-cfbce20d5f5a","Type":"ContainerDied","Data":"e7373bdb81f9b50ba8bbb0c9c25a406699b3a873853c84bc21738d29b87fdd5f"} Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.268577 4912 scope.go:117] "RemoveContainer" containerID="c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.283086 4912 scope.go:117] "RemoveContainer" containerID="cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.298248 4912 scope.go:117] "RemoveContainer" containerID="3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.325961 4912 scope.go:117] "RemoveContainer" containerID="c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf" Dec 03 00:40:05 crc kubenswrapper[4912]: E1203 00:40:05.326393 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf\": container with ID starting with c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf not found: ID does not exist" containerID="c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.326425 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf"} err="failed to get container status \"c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf\": rpc error: code = NotFound desc = could not find container \"c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf\": container with ID starting with c11a5eb41e587c10c90a0d97a8eea6fe95cf34e70b4542760e1dc9f1dacbbbbf not found: ID does not exist" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.326466 4912 scope.go:117] "RemoveContainer" containerID="cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461" Dec 03 00:40:05 crc kubenswrapper[4912]: E1203 00:40:05.326910 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461\": container with ID starting with cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461 not found: ID does not exist" containerID="cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.326966 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461"} err="failed to get container status \"cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461\": rpc error: code = NotFound desc = could not find container \"cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461\": container with ID starting with cec0b5f8872eeb4f8bb94781399f9d826af68e1d0551aaf869cb514b9f18b461 not found: ID does not exist" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.327001 4912 scope.go:117] "RemoveContainer" containerID="3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e" Dec 03 00:40:05 crc kubenswrapper[4912]: E1203 00:40:05.327328 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e\": container with ID starting with 3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e not found: ID does not exist" containerID="3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.327355 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e"} err="failed to get container status \"3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e\": rpc error: code = NotFound desc = could not find container \"3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e\": container with ID starting with 3cbb9794e2f2b5d43718cfb15af879952b68dcb86c8e0167154525afe999c74e not found: ID does not exist" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.372160 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-utilities\") pod \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.372294 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrkp8\" (UniqueName: \"kubernetes.io/projected/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-kube-api-access-mrkp8\") pod \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.372322 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-catalog-content\") pod \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\" (UID: \"f4577ec2-7bda-492f-8b34-cfbce20d5f5a\") " Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.373399 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-utilities" (OuterVolumeSpecName: "utilities") pod "f4577ec2-7bda-492f-8b34-cfbce20d5f5a" (UID: "f4577ec2-7bda-492f-8b34-cfbce20d5f5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.378688 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-kube-api-access-mrkp8" (OuterVolumeSpecName: "kube-api-access-mrkp8") pod "f4577ec2-7bda-492f-8b34-cfbce20d5f5a" (UID: "f4577ec2-7bda-492f-8b34-cfbce20d5f5a"). InnerVolumeSpecName "kube-api-access-mrkp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.390956 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4577ec2-7bda-492f-8b34-cfbce20d5f5a" (UID: "f4577ec2-7bda-492f-8b34-cfbce20d5f5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.474607 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.474641 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrkp8\" (UniqueName: \"kubernetes.io/projected/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-kube-api-access-mrkp8\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.474653 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4577ec2-7bda-492f-8b34-cfbce20d5f5a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.601255 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mmg5"] Dec 03 00:40:05 crc kubenswrapper[4912]: I1203 00:40:05.605593 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mmg5"] Dec 03 00:40:06 crc kubenswrapper[4912]: I1203 00:40:06.579808 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" path="/var/lib/kubelet/pods/f4577ec2-7bda-492f-8b34-cfbce20d5f5a/volumes" Dec 03 00:40:14 crc kubenswrapper[4912]: I1203 00:40:14.182298 4912 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 03 00:40:14 crc kubenswrapper[4912]: I1203 00:40:14.182660 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="33e70879-0df5-4529-969f-e88309e22cb0" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 00:40:18 crc kubenswrapper[4912]: I1203 00:40:18.077870 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:40:18 crc kubenswrapper[4912]: I1203 00:40:18.077941 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:40:18 crc kubenswrapper[4912]: I1203 00:40:18.077988 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:40:18 crc kubenswrapper[4912]: I1203 00:40:18.078671 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8ef572750b10ee1c03bff18588f184d2f43af325eeff79f4d4e863d818a97628"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:40:18 crc kubenswrapper[4912]: I1203 00:40:18.078737 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://8ef572750b10ee1c03bff18588f184d2f43af325eeff79f4d4e863d818a97628" gracePeriod=600 Dec 03 00:40:18 crc kubenswrapper[4912]: I1203 00:40:18.364368 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="8ef572750b10ee1c03bff18588f184d2f43af325eeff79f4d4e863d818a97628" exitCode=0 Dec 03 00:40:18 crc kubenswrapper[4912]: I1203 00:40:18.364602 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"8ef572750b10ee1c03bff18588f184d2f43af325eeff79f4d4e863d818a97628"} Dec 03 00:40:18 crc kubenswrapper[4912]: I1203 00:40:18.365025 4912 scope.go:117] "RemoveContainer" containerID="6db36669257e6a0bbce38d40b2b10188b2e1ca05791ecba752a07c893daa324b" Dec 03 00:40:19 crc kubenswrapper[4912]: I1203 00:40:19.374986 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"98a3aa8d06bd9638c9ea09b90a6f183e2b398dcf19a529ac1c64d4b6190f95cf"} Dec 03 00:40:24 crc kubenswrapper[4912]: I1203 00:40:24.183655 4912 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 03 00:40:24 crc kubenswrapper[4912]: I1203 00:40:24.184175 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="33e70879-0df5-4529-969f-e88309e22cb0" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 00:40:34 crc kubenswrapper[4912]: I1203 00:40:34.181413 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.002716 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-jzcxj"] Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.004157 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerName="extract-content" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.004183 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerName="extract-content" Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.004206 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerName="extract-utilities" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.004217 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerName="extract-utilities" Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.004241 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerName="extract-content" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.004254 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerName="extract-content" Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.004274 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerName="extract-utilities" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.004287 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerName="extract-utilities" Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.004310 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerName="registry-server" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.004321 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerName="registry-server" Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.004338 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerName="extract-content" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.004350 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerName="extract-content" Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.004375 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerName="extract-utilities" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.004386 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerName="extract-utilities" Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.004403 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerName="registry-server" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.004413 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerName="registry-server" Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.005522 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerName="registry-server" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.005540 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerName="registry-server" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.005770 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7d77826-4066-4f54-bdb9-e1fbbafdfa4c" containerName="registry-server" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.005792 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4577ec2-7bda-492f-8b34-cfbce20d5f5a" containerName="registry-server" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.005810 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="123154a6-35bb-4f9c-80c3-db4ef0fcdc82" containerName="registry-server" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.006792 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.016417 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.016902 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.017193 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-f7vmg" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.017467 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.018191 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.023315 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.028783 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-jzcxj"] Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.095900 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-jzcxj"] Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.096259 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-token\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.096722 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-sa-token\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.097078 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-syslog-receiver\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.097121 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0cce954e-b6f8-44e0-89ad-f7839a66ce09-datadir\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.097160 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-metrics\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: E1203 00:40:53.097238 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-pbjgd metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-jzcxj" podUID="0cce954e-b6f8-44e0-89ad-f7839a66ce09" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.097296 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-entrypoint\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.097483 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbjgd\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-kube-api-access-pbjgd\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.097626 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0cce954e-b6f8-44e0-89ad-f7839a66ce09-tmp\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.097742 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.097803 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config-openshift-service-cacrt\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.097929 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-trusted-ca\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.199969 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-trusted-ca\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200083 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-token\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200114 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-sa-token\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200151 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-syslog-receiver\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200181 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0cce954e-b6f8-44e0-89ad-f7839a66ce09-datadir\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200210 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-metrics\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200239 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-entrypoint\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200280 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbjgd\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-kube-api-access-pbjgd\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200309 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0cce954e-b6f8-44e0-89ad-f7839a66ce09-tmp\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200352 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.200395 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config-openshift-service-cacrt\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.201039 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-trusted-ca\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.201125 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config-openshift-service-cacrt\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.201749 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0cce954e-b6f8-44e0-89ad-f7839a66ce09-datadir\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.202011 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-entrypoint\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.202349 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.209497 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0cce954e-b6f8-44e0-89ad-f7839a66ce09-tmp\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.215612 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-syslog-receiver\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.221411 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-metrics\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.221846 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbjgd\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-kube-api-access-pbjgd\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.225907 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-token\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.226413 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-sa-token\") pod \"collector-jzcxj\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.599281 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.614278 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-jzcxj" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.708630 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-token\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.708846 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.708876 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-trusted-ca\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.708912 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbjgd\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-kube-api-access-pbjgd\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.708961 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0cce954e-b6f8-44e0-89ad-f7839a66ce09-datadir\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.708986 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-entrypoint\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.709007 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-metrics\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.709044 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config-openshift-service-cacrt\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.709069 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-syslog-receiver\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.709105 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0cce954e-b6f8-44e0-89ad-f7839a66ce09-tmp\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.709129 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-sa-token\") pod \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\" (UID: \"0cce954e-b6f8-44e0-89ad-f7839a66ce09\") " Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.709703 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.709950 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.709995 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cce954e-b6f8-44e0-89ad-f7839a66ce09-datadir" (OuterVolumeSpecName: "datadir") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.710267 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.710580 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config" (OuterVolumeSpecName: "config") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.716291 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-token" (OuterVolumeSpecName: "collector-token") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.718720 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-kube-api-access-pbjgd" (OuterVolumeSpecName: "kube-api-access-pbjgd") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "kube-api-access-pbjgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.720654 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.726753 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-sa-token" (OuterVolumeSpecName: "sa-token") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.732628 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cce954e-b6f8-44e0-89ad-f7839a66ce09-tmp" (OuterVolumeSpecName: "tmp") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.732643 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-metrics" (OuterVolumeSpecName: "metrics") pod "0cce954e-b6f8-44e0-89ad-f7839a66ce09" (UID: "0cce954e-b6f8-44e0-89ad-f7839a66ce09"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.811184 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.811612 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.811679 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbjgd\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-kube-api-access-pbjgd\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.811745 4912 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/0cce954e-b6f8-44e0-89ad-f7839a66ce09-datadir\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.811810 4912 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-entrypoint\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.811868 4912 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.811938 4912 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/0cce954e-b6f8-44e0-89ad-f7839a66ce09-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.812001 4912 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.812057 4912 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0cce954e-b6f8-44e0-89ad-f7839a66ce09-tmp\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.812119 4912 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/0cce954e-b6f8-44e0-89ad-f7839a66ce09-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:53 crc kubenswrapper[4912]: I1203 00:40:53.812177 4912 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/0cce954e-b6f8-44e0-89ad-f7839a66ce09-collector-token\") on node \"crc\" DevicePath \"\"" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.609323 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-jzcxj" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.656199 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-jzcxj"] Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.671085 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-jzcxj"] Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.675523 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-dccc6"] Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.676495 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.681172 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.681476 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.682377 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.690640 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-f7vmg" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.693687 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-dccc6"] Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.694221 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.696147 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.830832 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/5d5921d6-eb7c-4430-b71f-104dc4fdff22-collector-token\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831010 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/5d5921d6-eb7c-4430-b71f-104dc4fdff22-collector-syslog-receiver\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831077 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5d5921d6-eb7c-4430-b71f-104dc4fdff22-tmp\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831113 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-config-openshift-service-cacrt\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831156 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-trusted-ca\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831207 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-config\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831245 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/5d5921d6-eb7c-4430-b71f-104dc4fdff22-metrics\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831283 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/5d5921d6-eb7c-4430-b71f-104dc4fdff22-sa-token\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831314 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/5d5921d6-eb7c-4430-b71f-104dc4fdff22-datadir\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831493 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-entrypoint\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.831683 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2phts\" (UniqueName: \"kubernetes.io/projected/5d5921d6-eb7c-4430-b71f-104dc4fdff22-kube-api-access-2phts\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.933930 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-entrypoint\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.934131 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2phts\" (UniqueName: \"kubernetes.io/projected/5d5921d6-eb7c-4430-b71f-104dc4fdff22-kube-api-access-2phts\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.934182 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/5d5921d6-eb7c-4430-b71f-104dc4fdff22-collector-token\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.934242 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/5d5921d6-eb7c-4430-b71f-104dc4fdff22-collector-syslog-receiver\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.934284 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5d5921d6-eb7c-4430-b71f-104dc4fdff22-tmp\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.934314 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-config-openshift-service-cacrt\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.934351 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-trusted-ca\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.934390 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-config\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.934458 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/5d5921d6-eb7c-4430-b71f-104dc4fdff22-metrics\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.935564 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-entrypoint\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.935588 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-config-openshift-service-cacrt\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.935671 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/5d5921d6-eb7c-4430-b71f-104dc4fdff22-sa-token\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.935748 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/5d5921d6-eb7c-4430-b71f-104dc4fdff22-datadir\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.935873 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/5d5921d6-eb7c-4430-b71f-104dc4fdff22-datadir\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.936013 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-config\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.936332 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5d5921d6-eb7c-4430-b71f-104dc4fdff22-trusted-ca\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.938375 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/5d5921d6-eb7c-4430-b71f-104dc4fdff22-tmp\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.938705 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/5d5921d6-eb7c-4430-b71f-104dc4fdff22-metrics\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.938713 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/5d5921d6-eb7c-4430-b71f-104dc4fdff22-collector-token\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.953421 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/5d5921d6-eb7c-4430-b71f-104dc4fdff22-collector-syslog-receiver\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.954662 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/5d5921d6-eb7c-4430-b71f-104dc4fdff22-sa-token\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:54 crc kubenswrapper[4912]: I1203 00:40:54.958182 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2phts\" (UniqueName: \"kubernetes.io/projected/5d5921d6-eb7c-4430-b71f-104dc4fdff22-kube-api-access-2phts\") pod \"collector-dccc6\" (UID: \"5d5921d6-eb7c-4430-b71f-104dc4fdff22\") " pod="openshift-logging/collector-dccc6" Dec 03 00:40:55 crc kubenswrapper[4912]: I1203 00:40:55.030399 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-dccc6" Dec 03 00:40:55 crc kubenswrapper[4912]: I1203 00:40:55.471335 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-dccc6"] Dec 03 00:40:55 crc kubenswrapper[4912]: I1203 00:40:55.616329 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-dccc6" event={"ID":"5d5921d6-eb7c-4430-b71f-104dc4fdff22","Type":"ContainerStarted","Data":"d9d75bbc793018b4c2999a0224b9753aa53a562d2ce979056f996acba017d3d4"} Dec 03 00:40:56 crc kubenswrapper[4912]: I1203 00:40:56.583013 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cce954e-b6f8-44e0-89ad-f7839a66ce09" path="/var/lib/kubelet/pods/0cce954e-b6f8-44e0-89ad-f7839a66ce09/volumes" Dec 03 00:41:03 crc kubenswrapper[4912]: I1203 00:41:03.680896 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-dccc6" event={"ID":"5d5921d6-eb7c-4430-b71f-104dc4fdff22","Type":"ContainerStarted","Data":"c66d0e1024535339a4c87bdf72db494c75cd38ce457b85572a02517275a24d17"} Dec 03 00:41:03 crc kubenswrapper[4912]: I1203 00:41:03.790622 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-dccc6" podStartSLOduration=2.155977425 podStartE2EDuration="9.79060421s" podCreationTimestamp="2025-12-03 00:40:54 +0000 UTC" firstStartedPulling="2025-12-03 00:40:55.48763793 +0000 UTC m=+1041.129658510" lastFinishedPulling="2025-12-03 00:41:03.122264715 +0000 UTC m=+1048.764285295" observedRunningTime="2025-12-03 00:41:03.786144369 +0000 UTC m=+1049.428164939" watchObservedRunningTime="2025-12-03 00:41:03.79060421 +0000 UTC m=+1049.432624770" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.615787 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55"] Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.618182 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.621735 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.626298 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55"] Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.717964 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.718035 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.718132 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htq44\" (UniqueName: \"kubernetes.io/projected/e58f04a9-b86f-4988-98bb-b15d2b753399-kube-api-access-htq44\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.819238 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.819321 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htq44\" (UniqueName: \"kubernetes.io/projected/e58f04a9-b86f-4988-98bb-b15d2b753399-kube-api-access-htq44\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.819380 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.819885 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.820358 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.839528 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htq44\" (UniqueName: \"kubernetes.io/projected/e58f04a9-b86f-4988-98bb-b15d2b753399-kube-api-access-htq44\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:30 crc kubenswrapper[4912]: I1203 00:41:30.947759 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:31 crc kubenswrapper[4912]: I1203 00:41:31.588850 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55"] Dec 03 00:41:31 crc kubenswrapper[4912]: I1203 00:41:31.899747 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" event={"ID":"e58f04a9-b86f-4988-98bb-b15d2b753399","Type":"ContainerStarted","Data":"5d86045ded270a54059e916f6c639b11ea5b25183eb31e3db9b0ccb61c160160"} Dec 03 00:41:32 crc kubenswrapper[4912]: I1203 00:41:32.909994 4912 generic.go:334] "Generic (PLEG): container finished" podID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerID="57721006a46d5a02bf8e56da91617bc80d8c0402b3fb3972469ee24b14f84cba" exitCode=0 Dec 03 00:41:32 crc kubenswrapper[4912]: I1203 00:41:32.910053 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" event={"ID":"e58f04a9-b86f-4988-98bb-b15d2b753399","Type":"ContainerDied","Data":"57721006a46d5a02bf8e56da91617bc80d8c0402b3fb3972469ee24b14f84cba"} Dec 03 00:41:32 crc kubenswrapper[4912]: I1203 00:41:32.912126 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:41:34 crc kubenswrapper[4912]: I1203 00:41:34.927083 4912 generic.go:334] "Generic (PLEG): container finished" podID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerID="c39d6221e411c80f00959b7964ba91483c39eddc581bfb962a2073efd5eda287" exitCode=0 Dec 03 00:41:34 crc kubenswrapper[4912]: I1203 00:41:34.927401 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" event={"ID":"e58f04a9-b86f-4988-98bb-b15d2b753399","Type":"ContainerDied","Data":"c39d6221e411c80f00959b7964ba91483c39eddc581bfb962a2073efd5eda287"} Dec 03 00:41:35 crc kubenswrapper[4912]: I1203 00:41:35.942880 4912 generic.go:334] "Generic (PLEG): container finished" podID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerID="88452860dbe070d4f6f11e206a936dfcbb04aab1ffab2e48c8dd41f5660ec949" exitCode=0 Dec 03 00:41:35 crc kubenswrapper[4912]: I1203 00:41:35.942932 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" event={"ID":"e58f04a9-b86f-4988-98bb-b15d2b753399","Type":"ContainerDied","Data":"88452860dbe070d4f6f11e206a936dfcbb04aab1ffab2e48c8dd41f5660ec949"} Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.276581 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.319421 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-util\") pod \"e58f04a9-b86f-4988-98bb-b15d2b753399\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.319524 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-bundle\") pod \"e58f04a9-b86f-4988-98bb-b15d2b753399\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.319634 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htq44\" (UniqueName: \"kubernetes.io/projected/e58f04a9-b86f-4988-98bb-b15d2b753399-kube-api-access-htq44\") pod \"e58f04a9-b86f-4988-98bb-b15d2b753399\" (UID: \"e58f04a9-b86f-4988-98bb-b15d2b753399\") " Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.320329 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-bundle" (OuterVolumeSpecName: "bundle") pod "e58f04a9-b86f-4988-98bb-b15d2b753399" (UID: "e58f04a9-b86f-4988-98bb-b15d2b753399"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.328106 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e58f04a9-b86f-4988-98bb-b15d2b753399-kube-api-access-htq44" (OuterVolumeSpecName: "kube-api-access-htq44") pod "e58f04a9-b86f-4988-98bb-b15d2b753399" (UID: "e58f04a9-b86f-4988-98bb-b15d2b753399"). InnerVolumeSpecName "kube-api-access-htq44". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.421965 4912 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.422006 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htq44\" (UniqueName: \"kubernetes.io/projected/e58f04a9-b86f-4988-98bb-b15d2b753399-kube-api-access-htq44\") on node \"crc\" DevicePath \"\"" Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.423800 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-util" (OuterVolumeSpecName: "util") pod "e58f04a9-b86f-4988-98bb-b15d2b753399" (UID: "e58f04a9-b86f-4988-98bb-b15d2b753399"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.522782 4912 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e58f04a9-b86f-4988-98bb-b15d2b753399-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.959974 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" event={"ID":"e58f04a9-b86f-4988-98bb-b15d2b753399","Type":"ContainerDied","Data":"5d86045ded270a54059e916f6c639b11ea5b25183eb31e3db9b0ccb61c160160"} Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.960028 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d86045ded270a54059e916f6c639b11ea5b25183eb31e3db9b0ccb61c160160" Dec 03 00:41:37 crc kubenswrapper[4912]: I1203 00:41:37.960115 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55" Dec 03 00:41:38 crc kubenswrapper[4912]: E1203 00:41:38.073538 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode58f04a9_b86f_4988_98bb_b15d2b753399.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode58f04a9_b86f_4988_98bb_b15d2b753399.slice/crio-5d86045ded270a54059e916f6c639b11ea5b25183eb31e3db9b0ccb61c160160\": RecentStats: unable to find data in memory cache]" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.563591 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr"] Dec 03 00:41:40 crc kubenswrapper[4912]: E1203 00:41:40.564167 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerName="util" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.564180 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerName="util" Dec 03 00:41:40 crc kubenswrapper[4912]: E1203 00:41:40.564198 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerName="extract" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.564206 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerName="extract" Dec 03 00:41:40 crc kubenswrapper[4912]: E1203 00:41:40.564214 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerName="pull" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.564221 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerName="pull" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.564371 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="e58f04a9-b86f-4988-98bb-b15d2b753399" containerName="extract" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.564948 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.567720 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.569861 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bnp2p" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.570036 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.579734 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvql7\" (UniqueName: \"kubernetes.io/projected/48b78ad2-74d3-43ff-a69d-e729d3082199-kube-api-access-hvql7\") pod \"nmstate-operator-5b5b58f5c8-mxpqr\" (UID: \"48b78ad2-74d3-43ff-a69d-e729d3082199\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.590502 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr"] Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.681347 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvql7\" (UniqueName: \"kubernetes.io/projected/48b78ad2-74d3-43ff-a69d-e729d3082199-kube-api-access-hvql7\") pod \"nmstate-operator-5b5b58f5c8-mxpqr\" (UID: \"48b78ad2-74d3-43ff-a69d-e729d3082199\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.702763 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvql7\" (UniqueName: \"kubernetes.io/projected/48b78ad2-74d3-43ff-a69d-e729d3082199-kube-api-access-hvql7\") pod \"nmstate-operator-5b5b58f5c8-mxpqr\" (UID: \"48b78ad2-74d3-43ff-a69d-e729d3082199\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr" Dec 03 00:41:40 crc kubenswrapper[4912]: I1203 00:41:40.883936 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr" Dec 03 00:41:41 crc kubenswrapper[4912]: I1203 00:41:41.404136 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr"] Dec 03 00:41:41 crc kubenswrapper[4912]: I1203 00:41:41.987241 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr" event={"ID":"48b78ad2-74d3-43ff-a69d-e729d3082199","Type":"ContainerStarted","Data":"0fd10d8ae41cf77bb2aedd8ac0c77bafd6ec01142fe64d5db5c891cd5265f56f"} Dec 03 00:41:45 crc kubenswrapper[4912]: I1203 00:41:45.020537 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr" event={"ID":"48b78ad2-74d3-43ff-a69d-e729d3082199","Type":"ContainerStarted","Data":"94ab855f8ca92d4d22480895ba193bef4ea96448ca4ab3106961f628ee72aa4e"} Dec 03 00:41:45 crc kubenswrapper[4912]: I1203 00:41:45.040647 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-mxpqr" podStartSLOduration=1.809514954 podStartE2EDuration="5.040624216s" podCreationTimestamp="2025-12-03 00:41:40 +0000 UTC" firstStartedPulling="2025-12-03 00:41:41.413702607 +0000 UTC m=+1087.055723167" lastFinishedPulling="2025-12-03 00:41:44.644811869 +0000 UTC m=+1090.286832429" observedRunningTime="2025-12-03 00:41:45.03931952 +0000 UTC m=+1090.681340100" watchObservedRunningTime="2025-12-03 00:41:45.040624216 +0000 UTC m=+1090.682644786" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.318869 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw"] Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.321213 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.323865 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-6hv94" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.326597 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl"] Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.327842 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.330736 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.337114 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw"] Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.359012 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl"] Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.366075 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-667c7"] Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.367499 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.397393 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48kgj\" (UniqueName: \"kubernetes.io/projected/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-kube-api-access-48kgj\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.397478 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-ovs-socket\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.397506 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljrfp\" (UniqueName: \"kubernetes.io/projected/df373e95-5be1-45c8-a5a0-58dfd53caf3c-kube-api-access-ljrfp\") pod \"nmstate-webhook-5f6d4c5ccb-wt4wl\" (UID: \"df373e95-5be1-45c8-a5a0-58dfd53caf3c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.397537 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-dbus-socket\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.397558 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl6n4\" (UniqueName: \"kubernetes.io/projected/ff447976-7803-4a25-8491-f2094e165f86-kube-api-access-vl6n4\") pod \"nmstate-metrics-7f946cbc9-w95jw\" (UID: \"ff447976-7803-4a25-8491-f2094e165f86\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.397576 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-nmstate-lock\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.397643 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/df373e95-5be1-45c8-a5a0-58dfd53caf3c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-wt4wl\" (UID: \"df373e95-5be1-45c8-a5a0-58dfd53caf3c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.499507 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48kgj\" (UniqueName: \"kubernetes.io/projected/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-kube-api-access-48kgj\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.499583 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-ovs-socket\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.499613 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljrfp\" (UniqueName: \"kubernetes.io/projected/df373e95-5be1-45c8-a5a0-58dfd53caf3c-kube-api-access-ljrfp\") pod \"nmstate-webhook-5f6d4c5ccb-wt4wl\" (UID: \"df373e95-5be1-45c8-a5a0-58dfd53caf3c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.499648 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-dbus-socket\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.499674 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl6n4\" (UniqueName: \"kubernetes.io/projected/ff447976-7803-4a25-8491-f2094e165f86-kube-api-access-vl6n4\") pod \"nmstate-metrics-7f946cbc9-w95jw\" (UID: \"ff447976-7803-4a25-8491-f2094e165f86\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.499693 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-nmstate-lock\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.499754 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-ovs-socket\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.499975 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/df373e95-5be1-45c8-a5a0-58dfd53caf3c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-wt4wl\" (UID: \"df373e95-5be1-45c8-a5a0-58dfd53caf3c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.499986 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-dbus-socket\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.500055 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-nmstate-lock\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: E1203 00:41:50.500189 4912 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 03 00:41:50 crc kubenswrapper[4912]: E1203 00:41:50.500254 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df373e95-5be1-45c8-a5a0-58dfd53caf3c-tls-key-pair podName:df373e95-5be1-45c8-a5a0-58dfd53caf3c nodeName:}" failed. No retries permitted until 2025-12-03 00:41:51.000230288 +0000 UTC m=+1096.642250848 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/df373e95-5be1-45c8-a5a0-58dfd53caf3c-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-wt4wl" (UID: "df373e95-5be1-45c8-a5a0-58dfd53caf3c") : secret "openshift-nmstate-webhook" not found Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.501732 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc"] Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.502686 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.509741 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-f2hsv" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.511327 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.511526 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.527474 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48kgj\" (UniqueName: \"kubernetes.io/projected/af9b5997-afe3-45f8-8f6e-1b4b427c41cd-kube-api-access-48kgj\") pod \"nmstate-handler-667c7\" (UID: \"af9b5997-afe3-45f8-8f6e-1b4b427c41cd\") " pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.531806 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl6n4\" (UniqueName: \"kubernetes.io/projected/ff447976-7803-4a25-8491-f2094e165f86-kube-api-access-vl6n4\") pod \"nmstate-metrics-7f946cbc9-w95jw\" (UID: \"ff447976-7803-4a25-8491-f2094e165f86\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.534149 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc"] Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.540109 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljrfp\" (UniqueName: \"kubernetes.io/projected/df373e95-5be1-45c8-a5a0-58dfd53caf3c-kube-api-access-ljrfp\") pod \"nmstate-webhook-5f6d4c5ccb-wt4wl\" (UID: \"df373e95-5be1-45c8-a5a0-58dfd53caf3c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.642802 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c9a6fef-57f2-4add-99ff-2807455dd07d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.643337 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pnc2\" (UniqueName: \"kubernetes.io/projected/0c9a6fef-57f2-4add-99ff-2807455dd07d-kube-api-access-6pnc2\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.643624 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0c9a6fef-57f2-4add-99ff-2807455dd07d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.652742 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.687967 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.740593 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5674795f8c-jxb67"] Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.746192 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.747988 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c9a6fef-57f2-4add-99ff-2807455dd07d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.748043 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pnc2\" (UniqueName: \"kubernetes.io/projected/0c9a6fef-57f2-4add-99ff-2807455dd07d-kube-api-access-6pnc2\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.748124 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0c9a6fef-57f2-4add-99ff-2807455dd07d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.749211 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/0c9a6fef-57f2-4add-99ff-2807455dd07d-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:50 crc kubenswrapper[4912]: E1203 00:41:50.751682 4912 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 03 00:41:50 crc kubenswrapper[4912]: E1203 00:41:50.751867 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c9a6fef-57f2-4add-99ff-2807455dd07d-plugin-serving-cert podName:0c9a6fef-57f2-4add-99ff-2807455dd07d nodeName:}" failed. No retries permitted until 2025-12-03 00:41:51.251842289 +0000 UTC m=+1096.893862849 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/0c9a6fef-57f2-4add-99ff-2807455dd07d-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-7dtpc" (UID: "0c9a6fef-57f2-4add-99ff-2807455dd07d") : secret "plugin-serving-cert" not found Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.762416 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5674795f8c-jxb67"] Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.773500 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pnc2\" (UniqueName: \"kubernetes.io/projected/0c9a6fef-57f2-4add-99ff-2807455dd07d-kube-api-access-6pnc2\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.850353 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2s6l\" (UniqueName: \"kubernetes.io/projected/bcee1be8-8dde-47a2-85ec-69af93765fe9-kube-api-access-q2s6l\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.850450 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-service-ca\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.850515 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-oauth-serving-cert\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.850576 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-oauth-config\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.850641 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-trusted-ca-bundle\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.850669 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-config\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.850707 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-serving-cert\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.953471 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2s6l\" (UniqueName: \"kubernetes.io/projected/bcee1be8-8dde-47a2-85ec-69af93765fe9-kube-api-access-q2s6l\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.953550 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-service-ca\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.953594 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-oauth-serving-cert\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.953644 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-oauth-config\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.953671 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-trusted-ca-bundle\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.953694 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-config\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.953721 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-serving-cert\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.955722 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-service-ca\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.956831 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-trusted-ca-bundle\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.959279 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-config\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.960732 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-serving-cert\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.963266 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-oauth-config\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.972681 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-oauth-serving-cert\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:50 crc kubenswrapper[4912]: I1203 00:41:50.978565 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2s6l\" (UniqueName: \"kubernetes.io/projected/bcee1be8-8dde-47a2-85ec-69af93765fe9-kube-api-access-q2s6l\") pod \"console-5674795f8c-jxb67\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.057292 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/df373e95-5be1-45c8-a5a0-58dfd53caf3c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-wt4wl\" (UID: \"df373e95-5be1-45c8-a5a0-58dfd53caf3c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.062633 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/df373e95-5be1-45c8-a5a0-58dfd53caf3c-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-wt4wl\" (UID: \"df373e95-5be1-45c8-a5a0-58dfd53caf3c\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.067912 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-667c7" event={"ID":"af9b5997-afe3-45f8-8f6e-1b4b427c41cd","Type":"ContainerStarted","Data":"11a9a27ff84cdee76bd8265c99d064fbc82806f5482575c8fa9d050b8ee7469e"} Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.087517 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.175049 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw"] Dec 03 00:41:51 crc kubenswrapper[4912]: W1203 00:41:51.180833 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff447976_7803_4a25_8491_f2094e165f86.slice/crio-ef27ee85c719c291819bb9b1bef1d1f48f1f9c542f93434c4cd32891176051e6 WatchSource:0}: Error finding container ef27ee85c719c291819bb9b1bef1d1f48f1f9c542f93434c4cd32891176051e6: Status 404 returned error can't find the container with id ef27ee85c719c291819bb9b1bef1d1f48f1f9c542f93434c4cd32891176051e6 Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.260416 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c9a6fef-57f2-4add-99ff-2807455dd07d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.262854 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.267315 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c9a6fef-57f2-4add-99ff-2807455dd07d-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-7dtpc\" (UID: \"0c9a6fef-57f2-4add-99ff-2807455dd07d\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.419839 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.587880 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5674795f8c-jxb67"] Dec 03 00:41:51 crc kubenswrapper[4912]: W1203 00:41:51.594183 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcee1be8_8dde_47a2_85ec_69af93765fe9.slice/crio-2c3d574f42cdf24e131647b1cff087eb2e8a7e58bf26e15ce78cfd28b45f3b26 WatchSource:0}: Error finding container 2c3d574f42cdf24e131647b1cff087eb2e8a7e58bf26e15ce78cfd28b45f3b26: Status 404 returned error can't find the container with id 2c3d574f42cdf24e131647b1cff087eb2e8a7e58bf26e15ce78cfd28b45f3b26 Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.725527 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl"] Dec 03 00:41:51 crc kubenswrapper[4912]: W1203 00:41:51.750364 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf373e95_5be1_45c8_a5a0_58dfd53caf3c.slice/crio-839f35799ad1ded039ca060b10f1bfbd7c87d05528e4242ea9c39321abe76c2e WatchSource:0}: Error finding container 839f35799ad1ded039ca060b10f1bfbd7c87d05528e4242ea9c39321abe76c2e: Status 404 returned error can't find the container with id 839f35799ad1ded039ca060b10f1bfbd7c87d05528e4242ea9c39321abe76c2e Dec 03 00:41:51 crc kubenswrapper[4912]: I1203 00:41:51.806519 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc"] Dec 03 00:41:51 crc kubenswrapper[4912]: W1203 00:41:51.811951 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c9a6fef_57f2_4add_99ff_2807455dd07d.slice/crio-e9c066fba2c345b4fe56e38542b6f27fc8eb07b2b33c69ef290d14489923ba36 WatchSource:0}: Error finding container e9c066fba2c345b4fe56e38542b6f27fc8eb07b2b33c69ef290d14489923ba36: Status 404 returned error can't find the container with id e9c066fba2c345b4fe56e38542b6f27fc8eb07b2b33c69ef290d14489923ba36 Dec 03 00:41:52 crc kubenswrapper[4912]: I1203 00:41:52.076418 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw" event={"ID":"ff447976-7803-4a25-8491-f2094e165f86","Type":"ContainerStarted","Data":"ef27ee85c719c291819bb9b1bef1d1f48f1f9c542f93434c4cd32891176051e6"} Dec 03 00:41:52 crc kubenswrapper[4912]: I1203 00:41:52.077517 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" event={"ID":"0c9a6fef-57f2-4add-99ff-2807455dd07d","Type":"ContainerStarted","Data":"e9c066fba2c345b4fe56e38542b6f27fc8eb07b2b33c69ef290d14489923ba36"} Dec 03 00:41:52 crc kubenswrapper[4912]: I1203 00:41:52.078923 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5674795f8c-jxb67" event={"ID":"bcee1be8-8dde-47a2-85ec-69af93765fe9","Type":"ContainerStarted","Data":"abc6e08f599050799a8a4bad099d69cc8053722355adf6421ba2cea719bf3a84"} Dec 03 00:41:52 crc kubenswrapper[4912]: I1203 00:41:52.078957 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5674795f8c-jxb67" event={"ID":"bcee1be8-8dde-47a2-85ec-69af93765fe9","Type":"ContainerStarted","Data":"2c3d574f42cdf24e131647b1cff087eb2e8a7e58bf26e15ce78cfd28b45f3b26"} Dec 03 00:41:52 crc kubenswrapper[4912]: I1203 00:41:52.080102 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" event={"ID":"df373e95-5be1-45c8-a5a0-58dfd53caf3c","Type":"ContainerStarted","Data":"839f35799ad1ded039ca060b10f1bfbd7c87d05528e4242ea9c39321abe76c2e"} Dec 03 00:41:52 crc kubenswrapper[4912]: I1203 00:41:52.108849 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5674795f8c-jxb67" podStartSLOduration=2.1088239 podStartE2EDuration="2.1088239s" podCreationTimestamp="2025-12-03 00:41:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:41:52.098693184 +0000 UTC m=+1097.740713774" watchObservedRunningTime="2025-12-03 00:41:52.1088239 +0000 UTC m=+1097.750844460" Dec 03 00:41:54 crc kubenswrapper[4912]: I1203 00:41:54.097164 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" event={"ID":"df373e95-5be1-45c8-a5a0-58dfd53caf3c","Type":"ContainerStarted","Data":"f4d7b60688d83b3d4be2ee579c29a38fa613e84ac1f1b9ffaaa87348dee66d82"} Dec 03 00:41:54 crc kubenswrapper[4912]: I1203 00:41:54.098053 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:41:54 crc kubenswrapper[4912]: I1203 00:41:54.103810 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw" event={"ID":"ff447976-7803-4a25-8491-f2094e165f86","Type":"ContainerStarted","Data":"86a48c9968f43e1247ba56cf50ed812d5f1b59cba4029a7d69223d36984f552e"} Dec 03 00:41:54 crc kubenswrapper[4912]: I1203 00:41:54.122341 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" podStartSLOduration=2.022316751 podStartE2EDuration="4.122299454s" podCreationTimestamp="2025-12-03 00:41:50 +0000 UTC" firstStartedPulling="2025-12-03 00:41:51.752536907 +0000 UTC m=+1097.394557487" lastFinishedPulling="2025-12-03 00:41:53.85251963 +0000 UTC m=+1099.494540190" observedRunningTime="2025-12-03 00:41:54.115681174 +0000 UTC m=+1099.757701754" watchObservedRunningTime="2025-12-03 00:41:54.122299454 +0000 UTC m=+1099.764320034" Dec 03 00:41:55 crc kubenswrapper[4912]: I1203 00:41:55.111933 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-667c7" event={"ID":"af9b5997-afe3-45f8-8f6e-1b4b427c41cd","Type":"ContainerStarted","Data":"43dbf2f69d79e90b4b963a63deebbfda15b40aef86dd08924e745471cb991837"} Dec 03 00:41:55 crc kubenswrapper[4912]: I1203 00:41:55.112357 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:41:55 crc kubenswrapper[4912]: I1203 00:41:55.134424 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-667c7" podStartSLOduration=2.110949378 podStartE2EDuration="5.134403432s" podCreationTimestamp="2025-12-03 00:41:50 +0000 UTC" firstStartedPulling="2025-12-03 00:41:50.827945045 +0000 UTC m=+1096.469965605" lastFinishedPulling="2025-12-03 00:41:53.851399099 +0000 UTC m=+1099.493419659" observedRunningTime="2025-12-03 00:41:55.130534857 +0000 UTC m=+1100.772555437" watchObservedRunningTime="2025-12-03 00:41:55.134403432 +0000 UTC m=+1100.776424012" Dec 03 00:41:56 crc kubenswrapper[4912]: I1203 00:41:56.121903 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" event={"ID":"0c9a6fef-57f2-4add-99ff-2807455dd07d","Type":"ContainerStarted","Data":"53a30607efd3a59f0347a93e6b5734bbead7fe217d3a00557b7080988a1a7755"} Dec 03 00:41:56 crc kubenswrapper[4912]: I1203 00:41:56.142109 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-7dtpc" podStartSLOduration=2.479105122 podStartE2EDuration="6.142084209s" podCreationTimestamp="2025-12-03 00:41:50 +0000 UTC" firstStartedPulling="2025-12-03 00:41:51.814847379 +0000 UTC m=+1097.456867939" lastFinishedPulling="2025-12-03 00:41:55.477826456 +0000 UTC m=+1101.119847026" observedRunningTime="2025-12-03 00:41:56.139184981 +0000 UTC m=+1101.781205541" watchObservedRunningTime="2025-12-03 00:41:56.142084209 +0000 UTC m=+1101.784104769" Dec 03 00:41:57 crc kubenswrapper[4912]: I1203 00:41:57.130872 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw" event={"ID":"ff447976-7803-4a25-8491-f2094e165f86","Type":"ContainerStarted","Data":"0e60f4fbe4aa733cd4d47006655783cfc7a34c36cf65ee3fa2e6a212003d7020"} Dec 03 00:41:57 crc kubenswrapper[4912]: I1203 00:41:57.151944 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-w95jw" podStartSLOduration=1.675647731 podStartE2EDuration="7.151872355s" podCreationTimestamp="2025-12-03 00:41:50 +0000 UTC" firstStartedPulling="2025-12-03 00:41:51.184572048 +0000 UTC m=+1096.826592608" lastFinishedPulling="2025-12-03 00:41:56.660796672 +0000 UTC m=+1102.302817232" observedRunningTime="2025-12-03 00:41:57.146067477 +0000 UTC m=+1102.788088047" watchObservedRunningTime="2025-12-03 00:41:57.151872355 +0000 UTC m=+1102.793892915" Dec 03 00:42:00 crc kubenswrapper[4912]: I1203 00:42:00.716914 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-667c7" Dec 03 00:42:01 crc kubenswrapper[4912]: I1203 00:42:01.088667 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:42:01 crc kubenswrapper[4912]: I1203 00:42:01.088740 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:42:01 crc kubenswrapper[4912]: I1203 00:42:01.098357 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:42:01 crc kubenswrapper[4912]: I1203 00:42:01.166904 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:42:01 crc kubenswrapper[4912]: I1203 00:42:01.229924 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-cg7ls"] Dec 03 00:42:11 crc kubenswrapper[4912]: I1203 00:42:11.272787 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-wt4wl" Dec 03 00:42:18 crc kubenswrapper[4912]: I1203 00:42:18.078535 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:42:18 crc kubenswrapper[4912]: I1203 00:42:18.079651 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:42:26 crc kubenswrapper[4912]: I1203 00:42:26.273653 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-cg7ls" podUID="b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" containerName="console" containerID="cri-o://fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed" gracePeriod=15 Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.231932 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-cg7ls_b4a0b9b8-440c-4ac5-9e27-df5d838dfe93/console/0.log" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.232464 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.378843 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-cg7ls_b4a0b9b8-440c-4ac5-9e27-df5d838dfe93/console/0.log" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.378907 4912 generic.go:334] "Generic (PLEG): container finished" podID="b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" containerID="fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed" exitCode=2 Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.378935 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cg7ls" event={"ID":"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93","Type":"ContainerDied","Data":"fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed"} Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.378960 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-cg7ls" event={"ID":"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93","Type":"ContainerDied","Data":"3ced7fb72c9da7aafb7b0b7e960dd14acfb8a96219f5081e6832527e331b99dc"} Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.378979 4912 scope.go:117] "RemoveContainer" containerID="fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.379094 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-cg7ls" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.401028 4912 scope.go:117] "RemoveContainer" containerID="fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed" Dec 03 00:42:27 crc kubenswrapper[4912]: E1203 00:42:27.402029 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed\": container with ID starting with fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed not found: ID does not exist" containerID="fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.402072 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed"} err="failed to get container status \"fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed\": rpc error: code = NotFound desc = could not find container \"fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed\": container with ID starting with fcc2102613a24d8db9e935369ba4df863c96a55c889d30ed5ab5ace33e1151ed not found: ID does not exist" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.420048 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77zl8\" (UniqueName: \"kubernetes.io/projected/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-kube-api-access-77zl8\") pod \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.420111 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-service-ca\") pod \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.420132 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-serving-cert\") pod \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.420177 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-trusted-ca-bundle\") pod \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.420318 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-config\") pod \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.420336 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-oauth-config\") pod \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.420383 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-oauth-serving-cert\") pod \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\" (UID: \"b4a0b9b8-440c-4ac5-9e27-df5d838dfe93\") " Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.422241 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" (UID: "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.422231 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-service-ca" (OuterVolumeSpecName: "service-ca") pod "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" (UID: "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.422352 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-config" (OuterVolumeSpecName: "console-config") pod "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" (UID: "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.423114 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" (UID: "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.430806 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" (UID: "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.431214 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-kube-api-access-77zl8" (OuterVolumeSpecName: "kube-api-access-77zl8") pod "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" (UID: "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93"). InnerVolumeSpecName "kube-api-access-77zl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.431239 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" (UID: "b4a0b9b8-440c-4ac5-9e27-df5d838dfe93"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.524397 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.524466 4912 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.524482 4912 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.524498 4912 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.524513 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77zl8\" (UniqueName: \"kubernetes.io/projected/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-kube-api-access-77zl8\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.524527 4912 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.524539 4912 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.715380 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-cg7ls"] Dec 03 00:42:27 crc kubenswrapper[4912]: I1203 00:42:27.723367 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-cg7ls"] Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.583356 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" path="/var/lib/kubelet/pods/b4a0b9b8-440c-4ac5-9e27-df5d838dfe93/volumes" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.704159 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k"] Dec 03 00:42:28 crc kubenswrapper[4912]: E1203 00:42:28.704466 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" containerName="console" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.704477 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" containerName="console" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.704603 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4a0b9b8-440c-4ac5-9e27-df5d838dfe93" containerName="console" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.705539 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.708060 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.716084 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k"] Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.754132 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.754184 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x6lb\" (UniqueName: \"kubernetes.io/projected/8e6614d3-319c-4f5f-8937-fbe619a1c55a-kube-api-access-5x6lb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.754266 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.855457 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.855531 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.855573 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x6lb\" (UniqueName: \"kubernetes.io/projected/8e6614d3-319c-4f5f-8937-fbe619a1c55a-kube-api-access-5x6lb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.856250 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.856424 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:28 crc kubenswrapper[4912]: I1203 00:42:28.874807 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x6lb\" (UniqueName: \"kubernetes.io/projected/8e6614d3-319c-4f5f-8937-fbe619a1c55a-kube-api-access-5x6lb\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:29 crc kubenswrapper[4912]: I1203 00:42:29.033014 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:29 crc kubenswrapper[4912]: I1203 00:42:29.809465 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k"] Dec 03 00:42:30 crc kubenswrapper[4912]: I1203 00:42:30.410147 4912 generic.go:334] "Generic (PLEG): container finished" podID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerID="3fcf0d4091437d31a43418142fda3015f33cb00c4b88a52ee5bfa934149de71b" exitCode=0 Dec 03 00:42:30 crc kubenswrapper[4912]: I1203 00:42:30.410296 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" event={"ID":"8e6614d3-319c-4f5f-8937-fbe619a1c55a","Type":"ContainerDied","Data":"3fcf0d4091437d31a43418142fda3015f33cb00c4b88a52ee5bfa934149de71b"} Dec 03 00:42:30 crc kubenswrapper[4912]: I1203 00:42:30.411309 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" event={"ID":"8e6614d3-319c-4f5f-8937-fbe619a1c55a","Type":"ContainerStarted","Data":"fd1c3068f9c48b4bf063459223e64a8a5f26b36ea9980b8f8532b6dd35fd24c7"} Dec 03 00:42:32 crc kubenswrapper[4912]: I1203 00:42:32.430341 4912 generic.go:334] "Generic (PLEG): container finished" podID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerID="a83c9f73b90288e62d19f9ba8006ec43fe0a93c49e6c2d1f68b5fdc1bd12b47f" exitCode=0 Dec 03 00:42:32 crc kubenswrapper[4912]: I1203 00:42:32.430494 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" event={"ID":"8e6614d3-319c-4f5f-8937-fbe619a1c55a","Type":"ContainerDied","Data":"a83c9f73b90288e62d19f9ba8006ec43fe0a93c49e6c2d1f68b5fdc1bd12b47f"} Dec 03 00:42:33 crc kubenswrapper[4912]: I1203 00:42:33.442862 4912 generic.go:334] "Generic (PLEG): container finished" podID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerID="a38fad646fa69833e3e36bde1887abf9ac17b4ca9e623efa8b504d00c3c5f40c" exitCode=0 Dec 03 00:42:33 crc kubenswrapper[4912]: I1203 00:42:33.443029 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" event={"ID":"8e6614d3-319c-4f5f-8937-fbe619a1c55a","Type":"ContainerDied","Data":"a38fad646fa69833e3e36bde1887abf9ac17b4ca9e623efa8b504d00c3c5f40c"} Dec 03 00:42:34 crc kubenswrapper[4912]: I1203 00:42:34.823274 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:34 crc kubenswrapper[4912]: I1203 00:42:34.957525 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-bundle\") pod \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " Dec 03 00:42:34 crc kubenswrapper[4912]: I1203 00:42:34.957614 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x6lb\" (UniqueName: \"kubernetes.io/projected/8e6614d3-319c-4f5f-8937-fbe619a1c55a-kube-api-access-5x6lb\") pod \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " Dec 03 00:42:34 crc kubenswrapper[4912]: I1203 00:42:34.957739 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-util\") pod \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\" (UID: \"8e6614d3-319c-4f5f-8937-fbe619a1c55a\") " Dec 03 00:42:34 crc kubenswrapper[4912]: I1203 00:42:34.958486 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-bundle" (OuterVolumeSpecName: "bundle") pod "8e6614d3-319c-4f5f-8937-fbe619a1c55a" (UID: "8e6614d3-319c-4f5f-8937-fbe619a1c55a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:42:34 crc kubenswrapper[4912]: I1203 00:42:34.963343 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6614d3-319c-4f5f-8937-fbe619a1c55a-kube-api-access-5x6lb" (OuterVolumeSpecName: "kube-api-access-5x6lb") pod "8e6614d3-319c-4f5f-8937-fbe619a1c55a" (UID: "8e6614d3-319c-4f5f-8937-fbe619a1c55a"). InnerVolumeSpecName "kube-api-access-5x6lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:42:34 crc kubenswrapper[4912]: I1203 00:42:34.977749 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-util" (OuterVolumeSpecName: "util") pod "8e6614d3-319c-4f5f-8937-fbe619a1c55a" (UID: "8e6614d3-319c-4f5f-8937-fbe619a1c55a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:42:35 crc kubenswrapper[4912]: I1203 00:42:35.059559 4912 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:35 crc kubenswrapper[4912]: I1203 00:42:35.059636 4912 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e6614d3-319c-4f5f-8937-fbe619a1c55a-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:35 crc kubenswrapper[4912]: I1203 00:42:35.059649 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x6lb\" (UniqueName: \"kubernetes.io/projected/8e6614d3-319c-4f5f-8937-fbe619a1c55a-kube-api-access-5x6lb\") on node \"crc\" DevicePath \"\"" Dec 03 00:42:35 crc kubenswrapper[4912]: I1203 00:42:35.458352 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" event={"ID":"8e6614d3-319c-4f5f-8937-fbe619a1c55a","Type":"ContainerDied","Data":"fd1c3068f9c48b4bf063459223e64a8a5f26b36ea9980b8f8532b6dd35fd24c7"} Dec 03 00:42:35 crc kubenswrapper[4912]: I1203 00:42:35.458693 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd1c3068f9c48b4bf063459223e64a8a5f26b36ea9980b8f8532b6dd35fd24c7" Dec 03 00:42:35 crc kubenswrapper[4912]: I1203 00:42:35.458451 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.771135 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6"] Dec 03 00:42:43 crc kubenswrapper[4912]: E1203 00:42:43.771922 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerName="extract" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.771934 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerName="extract" Dec 03 00:42:43 crc kubenswrapper[4912]: E1203 00:42:43.771946 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerName="util" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.771952 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerName="util" Dec 03 00:42:43 crc kubenswrapper[4912]: E1203 00:42:43.771970 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerName="pull" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.771977 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerName="pull" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.772099 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6614d3-319c-4f5f-8937-fbe619a1c55a" containerName="extract" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.772600 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.774759 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.775124 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.775591 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.775723 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.779156 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-85nhz" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.799918 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6"] Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.902775 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxqm8\" (UniqueName: \"kubernetes.io/projected/ae8cc701-2530-4ef7-bd0c-596cb88ffc3a-kube-api-access-lxqm8\") pod \"metallb-operator-controller-manager-76f7d69d98-sl5x6\" (UID: \"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a\") " pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.902973 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ae8cc701-2530-4ef7-bd0c-596cb88ffc3a-apiservice-cert\") pod \"metallb-operator-controller-manager-76f7d69d98-sl5x6\" (UID: \"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a\") " pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:43 crc kubenswrapper[4912]: I1203 00:42:43.903045 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ae8cc701-2530-4ef7-bd0c-596cb88ffc3a-webhook-cert\") pod \"metallb-operator-controller-manager-76f7d69d98-sl5x6\" (UID: \"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a\") " pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.004251 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ae8cc701-2530-4ef7-bd0c-596cb88ffc3a-webhook-cert\") pod \"metallb-operator-controller-manager-76f7d69d98-sl5x6\" (UID: \"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a\") " pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.004389 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxqm8\" (UniqueName: \"kubernetes.io/projected/ae8cc701-2530-4ef7-bd0c-596cb88ffc3a-kube-api-access-lxqm8\") pod \"metallb-operator-controller-manager-76f7d69d98-sl5x6\" (UID: \"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a\") " pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.004478 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ae8cc701-2530-4ef7-bd0c-596cb88ffc3a-apiservice-cert\") pod \"metallb-operator-controller-manager-76f7d69d98-sl5x6\" (UID: \"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a\") " pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.012217 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ae8cc701-2530-4ef7-bd0c-596cb88ffc3a-apiservice-cert\") pod \"metallb-operator-controller-manager-76f7d69d98-sl5x6\" (UID: \"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a\") " pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.016222 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ae8cc701-2530-4ef7-bd0c-596cb88ffc3a-webhook-cert\") pod \"metallb-operator-controller-manager-76f7d69d98-sl5x6\" (UID: \"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a\") " pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.023415 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxqm8\" (UniqueName: \"kubernetes.io/projected/ae8cc701-2530-4ef7-bd0c-596cb88ffc3a-kube-api-access-lxqm8\") pod \"metallb-operator-controller-manager-76f7d69d98-sl5x6\" (UID: \"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a\") " pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.091246 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.214377 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb"] Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.216364 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.234867 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb"] Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.235213 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.249826 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-x84vq" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.250783 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.355276 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cb334b7c-36eb-4863-a2ad-2f4f126a9354-apiservice-cert\") pod \"metallb-operator-webhook-server-58487cf9c5-5g5xb\" (UID: \"cb334b7c-36eb-4863-a2ad-2f4f126a9354\") " pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.355835 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sjpn\" (UniqueName: \"kubernetes.io/projected/cb334b7c-36eb-4863-a2ad-2f4f126a9354-kube-api-access-6sjpn\") pod \"metallb-operator-webhook-server-58487cf9c5-5g5xb\" (UID: \"cb334b7c-36eb-4863-a2ad-2f4f126a9354\") " pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.355915 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cb334b7c-36eb-4863-a2ad-2f4f126a9354-webhook-cert\") pod \"metallb-operator-webhook-server-58487cf9c5-5g5xb\" (UID: \"cb334b7c-36eb-4863-a2ad-2f4f126a9354\") " pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.458090 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cb334b7c-36eb-4863-a2ad-2f4f126a9354-apiservice-cert\") pod \"metallb-operator-webhook-server-58487cf9c5-5g5xb\" (UID: \"cb334b7c-36eb-4863-a2ad-2f4f126a9354\") " pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.458200 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sjpn\" (UniqueName: \"kubernetes.io/projected/cb334b7c-36eb-4863-a2ad-2f4f126a9354-kube-api-access-6sjpn\") pod \"metallb-operator-webhook-server-58487cf9c5-5g5xb\" (UID: \"cb334b7c-36eb-4863-a2ad-2f4f126a9354\") " pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.458243 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cb334b7c-36eb-4863-a2ad-2f4f126a9354-webhook-cert\") pod \"metallb-operator-webhook-server-58487cf9c5-5g5xb\" (UID: \"cb334b7c-36eb-4863-a2ad-2f4f126a9354\") " pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.484726 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cb334b7c-36eb-4863-a2ad-2f4f126a9354-webhook-cert\") pod \"metallb-operator-webhook-server-58487cf9c5-5g5xb\" (UID: \"cb334b7c-36eb-4863-a2ad-2f4f126a9354\") " pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.488985 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cb334b7c-36eb-4863-a2ad-2f4f126a9354-apiservice-cert\") pod \"metallb-operator-webhook-server-58487cf9c5-5g5xb\" (UID: \"cb334b7c-36eb-4863-a2ad-2f4f126a9354\") " pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.489807 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sjpn\" (UniqueName: \"kubernetes.io/projected/cb334b7c-36eb-4863-a2ad-2f4f126a9354-kube-api-access-6sjpn\") pod \"metallb-operator-webhook-server-58487cf9c5-5g5xb\" (UID: \"cb334b7c-36eb-4863-a2ad-2f4f126a9354\") " pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.541600 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:44 crc kubenswrapper[4912]: I1203 00:42:44.613988 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6"] Dec 03 00:42:44 crc kubenswrapper[4912]: W1203 00:42:44.626189 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae8cc701_2530_4ef7_bd0c_596cb88ffc3a.slice/crio-8a5ce655a7e0cd68d57ded70010e7307c4425356101a83e90e036202e1419476 WatchSource:0}: Error finding container 8a5ce655a7e0cd68d57ded70010e7307c4425356101a83e90e036202e1419476: Status 404 returned error can't find the container with id 8a5ce655a7e0cd68d57ded70010e7307c4425356101a83e90e036202e1419476 Dec 03 00:42:45 crc kubenswrapper[4912]: I1203 00:42:45.084858 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb"] Dec 03 00:42:45 crc kubenswrapper[4912]: W1203 00:42:45.086100 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb334b7c_36eb_4863_a2ad_2f4f126a9354.slice/crio-0ccb2fd533eb3f2dff6af5ba126dc4ef74a3d159f348b61b88dd3a6bd6a26fc3 WatchSource:0}: Error finding container 0ccb2fd533eb3f2dff6af5ba126dc4ef74a3d159f348b61b88dd3a6bd6a26fc3: Status 404 returned error can't find the container with id 0ccb2fd533eb3f2dff6af5ba126dc4ef74a3d159f348b61b88dd3a6bd6a26fc3 Dec 03 00:42:45 crc kubenswrapper[4912]: I1203 00:42:45.536896 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" event={"ID":"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a","Type":"ContainerStarted","Data":"8a5ce655a7e0cd68d57ded70010e7307c4425356101a83e90e036202e1419476"} Dec 03 00:42:45 crc kubenswrapper[4912]: I1203 00:42:45.538137 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" event={"ID":"cb334b7c-36eb-4863-a2ad-2f4f126a9354","Type":"ContainerStarted","Data":"0ccb2fd533eb3f2dff6af5ba126dc4ef74a3d159f348b61b88dd3a6bd6a26fc3"} Dec 03 00:42:48 crc kubenswrapper[4912]: I1203 00:42:48.078576 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:42:48 crc kubenswrapper[4912]: I1203 00:42:48.079012 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:42:54 crc kubenswrapper[4912]: I1203 00:42:54.721182 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" event={"ID":"ae8cc701-2530-4ef7-bd0c-596cb88ffc3a","Type":"ContainerStarted","Data":"f1389aa145968b04520a18ad649000c27bc479422c9be365a98798e36cacd5a4"} Dec 03 00:42:54 crc kubenswrapper[4912]: I1203 00:42:54.722840 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:42:54 crc kubenswrapper[4912]: I1203 00:42:54.724423 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" event={"ID":"cb334b7c-36eb-4863-a2ad-2f4f126a9354","Type":"ContainerStarted","Data":"13b5030fce31156b6b6581b4767d40fb97dc4ba1ea6ed0d6b909d8fcb037dd7c"} Dec 03 00:42:54 crc kubenswrapper[4912]: I1203 00:42:54.724860 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:42:54 crc kubenswrapper[4912]: I1203 00:42:54.747831 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" podStartSLOduration=2.584384669 podStartE2EDuration="11.74781424s" podCreationTimestamp="2025-12-03 00:42:43 +0000 UTC" firstStartedPulling="2025-12-03 00:42:44.645406117 +0000 UTC m=+1150.287426677" lastFinishedPulling="2025-12-03 00:42:53.808835678 +0000 UTC m=+1159.450856248" observedRunningTime="2025-12-03 00:42:54.743742569 +0000 UTC m=+1160.385763149" watchObservedRunningTime="2025-12-03 00:42:54.74781424 +0000 UTC m=+1160.389834800" Dec 03 00:42:54 crc kubenswrapper[4912]: I1203 00:42:54.767368 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" podStartSLOduration=2.023012318 podStartE2EDuration="10.76734034s" podCreationTimestamp="2025-12-03 00:42:44 +0000 UTC" firstStartedPulling="2025-12-03 00:42:45.089611437 +0000 UTC m=+1150.731631987" lastFinishedPulling="2025-12-03 00:42:53.833939419 +0000 UTC m=+1159.475960009" observedRunningTime="2025-12-03 00:42:54.764281807 +0000 UTC m=+1160.406302377" watchObservedRunningTime="2025-12-03 00:42:54.76734034 +0000 UTC m=+1160.409360900" Dec 03 00:43:04 crc kubenswrapper[4912]: I1203 00:43:04.547624 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-58487cf9c5-5g5xb" Dec 03 00:43:18 crc kubenswrapper[4912]: I1203 00:43:18.077750 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:43:18 crc kubenswrapper[4912]: I1203 00:43:18.078689 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:43:18 crc kubenswrapper[4912]: I1203 00:43:18.078789 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:43:18 crc kubenswrapper[4912]: I1203 00:43:18.079886 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"98a3aa8d06bd9638c9ea09b90a6f183e2b398dcf19a529ac1c64d4b6190f95cf"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:43:18 crc kubenswrapper[4912]: I1203 00:43:18.079996 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://98a3aa8d06bd9638c9ea09b90a6f183e2b398dcf19a529ac1c64d4b6190f95cf" gracePeriod=600 Dec 03 00:43:18 crc kubenswrapper[4912]: I1203 00:43:18.944624 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="98a3aa8d06bd9638c9ea09b90a6f183e2b398dcf19a529ac1c64d4b6190f95cf" exitCode=0 Dec 03 00:43:18 crc kubenswrapper[4912]: I1203 00:43:18.944709 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"98a3aa8d06bd9638c9ea09b90a6f183e2b398dcf19a529ac1c64d4b6190f95cf"} Dec 03 00:43:18 crc kubenswrapper[4912]: I1203 00:43:18.945274 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"5d52c01534a739fd57a8ce1b61cd4e126b48cef8755e1aae547590b36373a992"} Dec 03 00:43:18 crc kubenswrapper[4912]: I1203 00:43:18.945306 4912 scope.go:117] "RemoveContainer" containerID="8ef572750b10ee1c03bff18588f184d2f43af325eeff79f4d4e863d818a97628" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.095658 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-76f7d69d98-sl5x6" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.895837 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-8c4wr"] Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.905313 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2"] Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.906225 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.906910 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.914490 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-mgprt" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.914753 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.914883 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.915629 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.928263 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2"] Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.988724 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnbw7\" (UniqueName: \"kubernetes.io/projected/1ccecd71-076b-4f63-830a-ad1507c443b8-kube-api-access-xnbw7\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.988798 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f8130d5-88a6-4a62-a381-e98a3207d9b3-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-jlww2\" (UID: \"9f8130d5-88a6-4a62-a381-e98a3207d9b3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.988861 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4qdt\" (UniqueName: \"kubernetes.io/projected/9f8130d5-88a6-4a62-a381-e98a3207d9b3-kube-api-access-c4qdt\") pod \"frr-k8s-webhook-server-7fcb986d4-jlww2\" (UID: \"9f8130d5-88a6-4a62-a381-e98a3207d9b3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.988900 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-frr-conf\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.988940 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1ccecd71-076b-4f63-830a-ad1507c443b8-frr-startup\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.988992 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-metrics\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.989025 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-reloader\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.989064 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ccecd71-076b-4f63-830a-ad1507c443b8-metrics-certs\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:24 crc kubenswrapper[4912]: I1203 00:43:24.989092 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-frr-sockets\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.025926 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-9nt4p"] Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.027362 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.032705 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.033015 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.033382 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bzrsx" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.033872 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.054702 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-sqxw4"] Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.056611 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.062213 4912 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.109611 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-frr-sockets\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.109708 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnbw7\" (UniqueName: \"kubernetes.io/projected/1ccecd71-076b-4f63-830a-ad1507c443b8-kube-api-access-xnbw7\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.109756 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f8130d5-88a6-4a62-a381-e98a3207d9b3-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-jlww2\" (UID: \"9f8130d5-88a6-4a62-a381-e98a3207d9b3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.109807 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4qdt\" (UniqueName: \"kubernetes.io/projected/9f8130d5-88a6-4a62-a381-e98a3207d9b3-kube-api-access-c4qdt\") pod \"frr-k8s-webhook-server-7fcb986d4-jlww2\" (UID: \"9f8130d5-88a6-4a62-a381-e98a3207d9b3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.109841 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-frr-conf\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.109871 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1ccecd71-076b-4f63-830a-ad1507c443b8-frr-startup\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.109924 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-metrics\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.109959 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-reloader\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.109999 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ccecd71-076b-4f63-830a-ad1507c443b8-metrics-certs\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: E1203 00:43:25.112323 4912 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 03 00:43:25 crc kubenswrapper[4912]: E1203 00:43:25.112418 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9f8130d5-88a6-4a62-a381-e98a3207d9b3-cert podName:9f8130d5-88a6-4a62-a381-e98a3207d9b3 nodeName:}" failed. No retries permitted until 2025-12-03 00:43:25.612387634 +0000 UTC m=+1191.254408194 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9f8130d5-88a6-4a62-a381-e98a3207d9b3-cert") pod "frr-k8s-webhook-server-7fcb986d4-jlww2" (UID: "9f8130d5-88a6-4a62-a381-e98a3207d9b3") : secret "frr-k8s-webhook-server-cert" not found Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.113017 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-frr-sockets\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.113681 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-frr-conf\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.114829 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-reloader\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.115003 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1ccecd71-076b-4f63-830a-ad1507c443b8-frr-startup\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.129515 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-sqxw4"] Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.114423 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1ccecd71-076b-4f63-830a-ad1507c443b8-metrics\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.137015 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1ccecd71-076b-4f63-830a-ad1507c443b8-metrics-certs\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.138200 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnbw7\" (UniqueName: \"kubernetes.io/projected/1ccecd71-076b-4f63-830a-ad1507c443b8-kube-api-access-xnbw7\") pod \"frr-k8s-8c4wr\" (UID: \"1ccecd71-076b-4f63-830a-ad1507c443b8\") " pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.152445 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4qdt\" (UniqueName: \"kubernetes.io/projected/9f8130d5-88a6-4a62-a381-e98a3207d9b3-kube-api-access-c4qdt\") pod \"frr-k8s-webhook-server-7fcb986d4-jlww2\" (UID: \"9f8130d5-88a6-4a62-a381-e98a3207d9b3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.211145 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-metrics-certs\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.211204 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-metallb-excludel2\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.211233 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4d86772-d2f0-4a24-80e7-339798fc2d2c-cert\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.211258 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4d86772-d2f0-4a24-80e7-339798fc2d2c-metrics-certs\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.211408 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jjqb\" (UniqueName: \"kubernetes.io/projected/b4d86772-d2f0-4a24-80e7-339798fc2d2c-kube-api-access-8jjqb\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.211564 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-memberlist\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.211667 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmg6b\" (UniqueName: \"kubernetes.io/projected/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-kube-api-access-rmg6b\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.259838 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.313664 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-metallb-excludel2\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.313712 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4d86772-d2f0-4a24-80e7-339798fc2d2c-cert\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.313741 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4d86772-d2f0-4a24-80e7-339798fc2d2c-metrics-certs\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.313771 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jjqb\" (UniqueName: \"kubernetes.io/projected/b4d86772-d2f0-4a24-80e7-339798fc2d2c-kube-api-access-8jjqb\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.313805 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-memberlist\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.313858 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmg6b\" (UniqueName: \"kubernetes.io/projected/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-kube-api-access-rmg6b\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.313926 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-metrics-certs\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: E1203 00:43:25.314574 4912 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 00:43:25 crc kubenswrapper[4912]: E1203 00:43:25.314657 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-memberlist podName:bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36 nodeName:}" failed. No retries permitted until 2025-12-03 00:43:25.814636055 +0000 UTC m=+1191.456656615 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-memberlist") pod "speaker-9nt4p" (UID: "bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36") : secret "metallb-memberlist" not found Dec 03 00:43:25 crc kubenswrapper[4912]: E1203 00:43:25.315402 4912 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 03 00:43:25 crc kubenswrapper[4912]: E1203 00:43:25.315459 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4d86772-d2f0-4a24-80e7-339798fc2d2c-metrics-certs podName:b4d86772-d2f0-4a24-80e7-339798fc2d2c nodeName:}" failed. No retries permitted until 2025-12-03 00:43:25.815445707 +0000 UTC m=+1191.457466267 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4d86772-d2f0-4a24-80e7-339798fc2d2c-metrics-certs") pod "controller-f8648f98b-sqxw4" (UID: "b4d86772-d2f0-4a24-80e7-339798fc2d2c") : secret "controller-certs-secret" not found Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.315396 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-metallb-excludel2\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.318957 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-metrics-certs\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.319617 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b4d86772-d2f0-4a24-80e7-339798fc2d2c-cert\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.333065 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jjqb\" (UniqueName: \"kubernetes.io/projected/b4d86772-d2f0-4a24-80e7-339798fc2d2c-kube-api-access-8jjqb\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.335795 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmg6b\" (UniqueName: \"kubernetes.io/projected/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-kube-api-access-rmg6b\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.618326 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f8130d5-88a6-4a62-a381-e98a3207d9b3-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-jlww2\" (UID: \"9f8130d5-88a6-4a62-a381-e98a3207d9b3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.630301 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9f8130d5-88a6-4a62-a381-e98a3207d9b3-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-jlww2\" (UID: \"9f8130d5-88a6-4a62-a381-e98a3207d9b3\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.821800 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4d86772-d2f0-4a24-80e7-339798fc2d2c-metrics-certs\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.822138 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-memberlist\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:25 crc kubenswrapper[4912]: E1203 00:43:25.822377 4912 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 00:43:25 crc kubenswrapper[4912]: E1203 00:43:25.822544 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-memberlist podName:bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36 nodeName:}" failed. No retries permitted until 2025-12-03 00:43:26.822503493 +0000 UTC m=+1192.464524053 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-memberlist") pod "speaker-9nt4p" (UID: "bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36") : secret "metallb-memberlist" not found Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.829347 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4d86772-d2f0-4a24-80e7-339798fc2d2c-metrics-certs\") pod \"controller-f8648f98b-sqxw4\" (UID: \"b4d86772-d2f0-4a24-80e7-339798fc2d2c\") " pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.845389 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:25 crc kubenswrapper[4912]: I1203 00:43:25.978530 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:26 crc kubenswrapper[4912]: I1203 00:43:26.017749 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerStarted","Data":"799113506daf3426e9286c062dbb7997adeca64f8983353bea1d1a24f674550b"} Dec 03 00:43:26 crc kubenswrapper[4912]: I1203 00:43:26.277631 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-sqxw4"] Dec 03 00:43:26 crc kubenswrapper[4912]: I1203 00:43:26.383896 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2"] Dec 03 00:43:26 crc kubenswrapper[4912]: W1203 00:43:26.402145 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f8130d5_88a6_4a62_a381_e98a3207d9b3.slice/crio-fbfb81cf6f39731f5a7447a0f8328ab2f1f684de7f39a43d05d606cd8982418e WatchSource:0}: Error finding container fbfb81cf6f39731f5a7447a0f8328ab2f1f684de7f39a43d05d606cd8982418e: Status 404 returned error can't find the container with id fbfb81cf6f39731f5a7447a0f8328ab2f1f684de7f39a43d05d606cd8982418e Dec 03 00:43:26 crc kubenswrapper[4912]: I1203 00:43:26.844501 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-memberlist\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:26 crc kubenswrapper[4912]: I1203 00:43:26.854196 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36-memberlist\") pod \"speaker-9nt4p\" (UID: \"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36\") " pod="metallb-system/speaker-9nt4p" Dec 03 00:43:27 crc kubenswrapper[4912]: I1203 00:43:27.045063 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-sqxw4" event={"ID":"b4d86772-d2f0-4a24-80e7-339798fc2d2c","Type":"ContainerStarted","Data":"56d9421f8a82b00e3b8827b2f3e4445dbd0c425cd3cfa8fdba86ad58bbb434ed"} Dec 03 00:43:27 crc kubenswrapper[4912]: I1203 00:43:27.045130 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-sqxw4" event={"ID":"b4d86772-d2f0-4a24-80e7-339798fc2d2c","Type":"ContainerStarted","Data":"ea075a6079bc295d25a06b403fe84274f854aadae3cb76bdded3de2770f3b3e2"} Dec 03 00:43:27 crc kubenswrapper[4912]: I1203 00:43:27.045245 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-sqxw4" event={"ID":"b4d86772-d2f0-4a24-80e7-339798fc2d2c","Type":"ContainerStarted","Data":"ad7f2e8252b21fc4ac4bd6f40ece4390c739fe0254f7ae453776fe80b8f4ac55"} Dec 03 00:43:27 crc kubenswrapper[4912]: I1203 00:43:27.046712 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:27 crc kubenswrapper[4912]: I1203 00:43:27.053137 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" event={"ID":"9f8130d5-88a6-4a62-a381-e98a3207d9b3","Type":"ContainerStarted","Data":"fbfb81cf6f39731f5a7447a0f8328ab2f1f684de7f39a43d05d606cd8982418e"} Dec 03 00:43:27 crc kubenswrapper[4912]: I1203 00:43:27.070333 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-sqxw4" podStartSLOduration=2.07031191 podStartE2EDuration="2.07031191s" podCreationTimestamp="2025-12-03 00:43:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:43:27.067769452 +0000 UTC m=+1192.709790052" watchObservedRunningTime="2025-12-03 00:43:27.07031191 +0000 UTC m=+1192.712332480" Dec 03 00:43:27 crc kubenswrapper[4912]: I1203 00:43:27.152127 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9nt4p" Dec 03 00:43:27 crc kubenswrapper[4912]: W1203 00:43:27.180855 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd9f5e2d_164a_4a09_8cf4_57f1ecd84e36.slice/crio-7270a534a9b5c106fcafcb9e9e7936c6a0d8f29a6db96ef8be3c01df8049a424 WatchSource:0}: Error finding container 7270a534a9b5c106fcafcb9e9e7936c6a0d8f29a6db96ef8be3c01df8049a424: Status 404 returned error can't find the container with id 7270a534a9b5c106fcafcb9e9e7936c6a0d8f29a6db96ef8be3c01df8049a424 Dec 03 00:43:28 crc kubenswrapper[4912]: I1203 00:43:28.063572 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9nt4p" event={"ID":"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36","Type":"ContainerStarted","Data":"dfbfe6a1251cdffb19d6690b0df973c635dd65f368c919e78741cae38bb6c194"} Dec 03 00:43:28 crc kubenswrapper[4912]: I1203 00:43:28.063675 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9nt4p" event={"ID":"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36","Type":"ContainerStarted","Data":"5594752fbf66651a10904f5b336a2ca353eb371fdd68d152b9c280f2d48a7dad"} Dec 03 00:43:28 crc kubenswrapper[4912]: I1203 00:43:28.063686 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9nt4p" event={"ID":"bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36","Type":"ContainerStarted","Data":"7270a534a9b5c106fcafcb9e9e7936c6a0d8f29a6db96ef8be3c01df8049a424"} Dec 03 00:43:28 crc kubenswrapper[4912]: I1203 00:43:28.063974 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-9nt4p" Dec 03 00:43:28 crc kubenswrapper[4912]: I1203 00:43:28.089370 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-9nt4p" podStartSLOduration=4.089333856 podStartE2EDuration="4.089333856s" podCreationTimestamp="2025-12-03 00:43:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:43:28.083157569 +0000 UTC m=+1193.725178139" watchObservedRunningTime="2025-12-03 00:43:28.089333856 +0000 UTC m=+1193.731354416" Dec 03 00:43:35 crc kubenswrapper[4912]: I1203 00:43:35.133154 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" event={"ID":"9f8130d5-88a6-4a62-a381-e98a3207d9b3","Type":"ContainerStarted","Data":"6203d247bd73789ee53d8c4b9e10ca10062960309ae572ec931e4862fc56b0b3"} Dec 03 00:43:35 crc kubenswrapper[4912]: I1203 00:43:35.134177 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:35 crc kubenswrapper[4912]: I1203 00:43:35.137782 4912 generic.go:334] "Generic (PLEG): container finished" podID="1ccecd71-076b-4f63-830a-ad1507c443b8" containerID="8d28bb4f14abd64b3bb1044ecbf6d823db09953f6c032cbd814a89e0564b76c0" exitCode=0 Dec 03 00:43:35 crc kubenswrapper[4912]: I1203 00:43:35.137868 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerDied","Data":"8d28bb4f14abd64b3bb1044ecbf6d823db09953f6c032cbd814a89e0564b76c0"} Dec 03 00:43:35 crc kubenswrapper[4912]: I1203 00:43:35.158090 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" podStartSLOduration=3.455727195 podStartE2EDuration="11.158061367s" podCreationTimestamp="2025-12-03 00:43:24 +0000 UTC" firstStartedPulling="2025-12-03 00:43:26.406086028 +0000 UTC m=+1192.048106588" lastFinishedPulling="2025-12-03 00:43:34.10842019 +0000 UTC m=+1199.750440760" observedRunningTime="2025-12-03 00:43:35.155526328 +0000 UTC m=+1200.797546898" watchObservedRunningTime="2025-12-03 00:43:35.158061367 +0000 UTC m=+1200.800081927" Dec 03 00:43:36 crc kubenswrapper[4912]: I1203 00:43:36.147404 4912 generic.go:334] "Generic (PLEG): container finished" podID="1ccecd71-076b-4f63-830a-ad1507c443b8" containerID="5d827d6d2473898e827a6d0c8f0f54450fbd8b4c4fdb577b56746a7f651106d7" exitCode=0 Dec 03 00:43:36 crc kubenswrapper[4912]: I1203 00:43:36.147505 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerDied","Data":"5d827d6d2473898e827a6d0c8f0f54450fbd8b4c4fdb577b56746a7f651106d7"} Dec 03 00:43:37 crc kubenswrapper[4912]: I1203 00:43:37.162576 4912 generic.go:334] "Generic (PLEG): container finished" podID="1ccecd71-076b-4f63-830a-ad1507c443b8" containerID="9f843072876dc3a820584770887741f647e5dc8d113aa08ea6f70bddf5048b80" exitCode=0 Dec 03 00:43:37 crc kubenswrapper[4912]: I1203 00:43:37.162708 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerDied","Data":"9f843072876dc3a820584770887741f647e5dc8d113aa08ea6f70bddf5048b80"} Dec 03 00:43:37 crc kubenswrapper[4912]: I1203 00:43:37.194358 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-9nt4p" Dec 03 00:43:38 crc kubenswrapper[4912]: I1203 00:43:38.185043 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerStarted","Data":"bb89fcddf61301158ed21142278964a7624db55586c01575620ec63b4efff51e"} Dec 03 00:43:38 crc kubenswrapper[4912]: I1203 00:43:38.185664 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerStarted","Data":"7e2af5082b8dc61ceb846cb42448337418257e091a8b3644bf5413930bfc033f"} Dec 03 00:43:38 crc kubenswrapper[4912]: I1203 00:43:38.185700 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerStarted","Data":"2cf779d58a2ec8834c86bcdd5f20b25e2951c004522d944cf783cc29c71dffdf"} Dec 03 00:43:38 crc kubenswrapper[4912]: I1203 00:43:38.185740 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerStarted","Data":"32cade8e28b7b2c612b38d3a2b6c76eac55502a36fd5de602ca658f5905935c0"} Dec 03 00:43:39 crc kubenswrapper[4912]: I1203 00:43:39.217534 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerStarted","Data":"bfbaaab92f6ba1f634d81a399f7a1b04e122adbb998af05c0a74519a95003a45"} Dec 03 00:43:39 crc kubenswrapper[4912]: I1203 00:43:39.218004 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8c4wr" event={"ID":"1ccecd71-076b-4f63-830a-ad1507c443b8","Type":"ContainerStarted","Data":"0202c0fb79becc904017c776214e42cf5d03b1412c1f87d419dae37c7644d514"} Dec 03 00:43:39 crc kubenswrapper[4912]: I1203 00:43:39.218138 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:39 crc kubenswrapper[4912]: I1203 00:43:39.248417 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-8c4wr" podStartSLOduration=6.643130999 podStartE2EDuration="15.248399496s" podCreationTimestamp="2025-12-03 00:43:24 +0000 UTC" firstStartedPulling="2025-12-03 00:43:25.471776512 +0000 UTC m=+1191.113797072" lastFinishedPulling="2025-12-03 00:43:34.077044999 +0000 UTC m=+1199.719065569" observedRunningTime="2025-12-03 00:43:39.245868308 +0000 UTC m=+1204.887888888" watchObservedRunningTime="2025-12-03 00:43:39.248399496 +0000 UTC m=+1204.890420056" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.261654 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.310349 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.668518 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tjdj9"] Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.670130 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tjdj9" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.672947 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-mw4dk" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.677368 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.680419 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tjdj9"] Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.683780 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.838609 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d8mt\" (UniqueName: \"kubernetes.io/projected/594281e0-91bf-4ca1-8937-b379f396d46d-kube-api-access-5d8mt\") pod \"openstack-operator-index-tjdj9\" (UID: \"594281e0-91bf-4ca1-8937-b379f396d46d\") " pod="openstack-operators/openstack-operator-index-tjdj9" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.940375 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d8mt\" (UniqueName: \"kubernetes.io/projected/594281e0-91bf-4ca1-8937-b379f396d46d-kube-api-access-5d8mt\") pod \"openstack-operator-index-tjdj9\" (UID: \"594281e0-91bf-4ca1-8937-b379f396d46d\") " pod="openstack-operators/openstack-operator-index-tjdj9" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.965129 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d8mt\" (UniqueName: \"kubernetes.io/projected/594281e0-91bf-4ca1-8937-b379f396d46d-kube-api-access-5d8mt\") pod \"openstack-operator-index-tjdj9\" (UID: \"594281e0-91bf-4ca1-8937-b379f396d46d\") " pod="openstack-operators/openstack-operator-index-tjdj9" Dec 03 00:43:40 crc kubenswrapper[4912]: I1203 00:43:40.994968 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tjdj9" Dec 03 00:43:41 crc kubenswrapper[4912]: I1203 00:43:41.474186 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tjdj9"] Dec 03 00:43:42 crc kubenswrapper[4912]: I1203 00:43:42.246109 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tjdj9" event={"ID":"594281e0-91bf-4ca1-8937-b379f396d46d","Type":"ContainerStarted","Data":"3bda02e018afa72ac3a8fade552a69ba0f674b7b01494c0c97de99b4b90ddcb6"} Dec 03 00:43:43 crc kubenswrapper[4912]: I1203 00:43:43.634902 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-tjdj9"] Dec 03 00:43:44 crc kubenswrapper[4912]: I1203 00:43:44.264351 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-4nzrg"] Dec 03 00:43:44 crc kubenswrapper[4912]: I1203 00:43:44.265501 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4nzrg" Dec 03 00:43:44 crc kubenswrapper[4912]: I1203 00:43:44.281115 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4nzrg"] Dec 03 00:43:44 crc kubenswrapper[4912]: I1203 00:43:44.416966 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxnml\" (UniqueName: \"kubernetes.io/projected/5d286a0d-a818-4cb6-a798-135fda9d947e-kube-api-access-jxnml\") pod \"openstack-operator-index-4nzrg\" (UID: \"5d286a0d-a818-4cb6-a798-135fda9d947e\") " pod="openstack-operators/openstack-operator-index-4nzrg" Dec 03 00:43:44 crc kubenswrapper[4912]: I1203 00:43:44.518790 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxnml\" (UniqueName: \"kubernetes.io/projected/5d286a0d-a818-4cb6-a798-135fda9d947e-kube-api-access-jxnml\") pod \"openstack-operator-index-4nzrg\" (UID: \"5d286a0d-a818-4cb6-a798-135fda9d947e\") " pod="openstack-operators/openstack-operator-index-4nzrg" Dec 03 00:43:44 crc kubenswrapper[4912]: I1203 00:43:44.560518 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxnml\" (UniqueName: \"kubernetes.io/projected/5d286a0d-a818-4cb6-a798-135fda9d947e-kube-api-access-jxnml\") pod \"openstack-operator-index-4nzrg\" (UID: \"5d286a0d-a818-4cb6-a798-135fda9d947e\") " pod="openstack-operators/openstack-operator-index-4nzrg" Dec 03 00:43:44 crc kubenswrapper[4912]: I1203 00:43:44.626890 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4nzrg" Dec 03 00:43:45 crc kubenswrapper[4912]: I1203 00:43:45.444820 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4nzrg"] Dec 03 00:43:45 crc kubenswrapper[4912]: W1203 00:43:45.445622 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d286a0d_a818_4cb6_a798_135fda9d947e.slice/crio-4be90f261eaada23b9386d968ed8658ddedc7217ccf4b3c7fbf94fda7b554b0b WatchSource:0}: Error finding container 4be90f261eaada23b9386d968ed8658ddedc7217ccf4b3c7fbf94fda7b554b0b: Status 404 returned error can't find the container with id 4be90f261eaada23b9386d968ed8658ddedc7217ccf4b3c7fbf94fda7b554b0b Dec 03 00:43:45 crc kubenswrapper[4912]: I1203 00:43:45.857098 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-jlww2" Dec 03 00:43:45 crc kubenswrapper[4912]: I1203 00:43:45.986555 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-sqxw4" Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.295836 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4nzrg" event={"ID":"5d286a0d-a818-4cb6-a798-135fda9d947e","Type":"ContainerStarted","Data":"38553378d24f589e47bb032b2785a211433b30e3143db56fa6d800b2cc26b806"} Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.295881 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4nzrg" event={"ID":"5d286a0d-a818-4cb6-a798-135fda9d947e","Type":"ContainerStarted","Data":"4be90f261eaada23b9386d968ed8658ddedc7217ccf4b3c7fbf94fda7b554b0b"} Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.298333 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tjdj9" event={"ID":"594281e0-91bf-4ca1-8937-b379f396d46d","Type":"ContainerStarted","Data":"d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0"} Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.298448 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-tjdj9" podUID="594281e0-91bf-4ca1-8937-b379f396d46d" containerName="registry-server" containerID="cri-o://d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0" gracePeriod=2 Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.317533 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-4nzrg" podStartSLOduration=2.245294805 podStartE2EDuration="2.317509136s" podCreationTimestamp="2025-12-03 00:43:44 +0000 UTC" firstStartedPulling="2025-12-03 00:43:45.449423848 +0000 UTC m=+1211.091482619" lastFinishedPulling="2025-12-03 00:43:45.52167639 +0000 UTC m=+1211.163696950" observedRunningTime="2025-12-03 00:43:46.31138626 +0000 UTC m=+1211.953406840" watchObservedRunningTime="2025-12-03 00:43:46.317509136 +0000 UTC m=+1211.959529696" Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.339325 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tjdj9" podStartSLOduration=2.645108512 podStartE2EDuration="6.339307127s" podCreationTimestamp="2025-12-03 00:43:40 +0000 UTC" firstStartedPulling="2025-12-03 00:43:41.49470802 +0000 UTC m=+1207.136728580" lastFinishedPulling="2025-12-03 00:43:45.188906615 +0000 UTC m=+1210.830927195" observedRunningTime="2025-12-03 00:43:46.332423591 +0000 UTC m=+1211.974444171" watchObservedRunningTime="2025-12-03 00:43:46.339307127 +0000 UTC m=+1211.981327687" Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.762077 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tjdj9" Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.859775 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d8mt\" (UniqueName: \"kubernetes.io/projected/594281e0-91bf-4ca1-8937-b379f396d46d-kube-api-access-5d8mt\") pod \"594281e0-91bf-4ca1-8937-b379f396d46d\" (UID: \"594281e0-91bf-4ca1-8937-b379f396d46d\") " Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.867589 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/594281e0-91bf-4ca1-8937-b379f396d46d-kube-api-access-5d8mt" (OuterVolumeSpecName: "kube-api-access-5d8mt") pod "594281e0-91bf-4ca1-8937-b379f396d46d" (UID: "594281e0-91bf-4ca1-8937-b379f396d46d"). InnerVolumeSpecName "kube-api-access-5d8mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:43:46 crc kubenswrapper[4912]: I1203 00:43:46.962958 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d8mt\" (UniqueName: \"kubernetes.io/projected/594281e0-91bf-4ca1-8937-b379f396d46d-kube-api-access-5d8mt\") on node \"crc\" DevicePath \"\"" Dec 03 00:43:47 crc kubenswrapper[4912]: I1203 00:43:47.306661 4912 generic.go:334] "Generic (PLEG): container finished" podID="594281e0-91bf-4ca1-8937-b379f396d46d" containerID="d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0" exitCode=0 Dec 03 00:43:47 crc kubenswrapper[4912]: I1203 00:43:47.306727 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tjdj9" event={"ID":"594281e0-91bf-4ca1-8937-b379f396d46d","Type":"ContainerDied","Data":"d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0"} Dec 03 00:43:47 crc kubenswrapper[4912]: I1203 00:43:47.306786 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tjdj9" event={"ID":"594281e0-91bf-4ca1-8937-b379f396d46d","Type":"ContainerDied","Data":"3bda02e018afa72ac3a8fade552a69ba0f674b7b01494c0c97de99b4b90ddcb6"} Dec 03 00:43:47 crc kubenswrapper[4912]: I1203 00:43:47.306806 4912 scope.go:117] "RemoveContainer" containerID="d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0" Dec 03 00:43:47 crc kubenswrapper[4912]: I1203 00:43:47.306742 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tjdj9" Dec 03 00:43:47 crc kubenswrapper[4912]: I1203 00:43:47.346335 4912 scope.go:117] "RemoveContainer" containerID="d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0" Dec 03 00:43:47 crc kubenswrapper[4912]: E1203 00:43:47.346898 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0\": container with ID starting with d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0 not found: ID does not exist" containerID="d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0" Dec 03 00:43:47 crc kubenswrapper[4912]: I1203 00:43:47.346934 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0"} err="failed to get container status \"d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0\": rpc error: code = NotFound desc = could not find container \"d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0\": container with ID starting with d6b1434f63f65865bd0eb1acef96c7f8956d8397f13b363c43713d72326e2ab0 not found: ID does not exist" Dec 03 00:43:47 crc kubenswrapper[4912]: I1203 00:43:47.349768 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-tjdj9"] Dec 03 00:43:47 crc kubenswrapper[4912]: I1203 00:43:47.356205 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-tjdj9"] Dec 03 00:43:48 crc kubenswrapper[4912]: I1203 00:43:48.580875 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="594281e0-91bf-4ca1-8937-b379f396d46d" path="/var/lib/kubelet/pods/594281e0-91bf-4ca1-8937-b379f396d46d/volumes" Dec 03 00:43:54 crc kubenswrapper[4912]: I1203 00:43:54.628465 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-4nzrg" Dec 03 00:43:54 crc kubenswrapper[4912]: I1203 00:43:54.628852 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-4nzrg" Dec 03 00:43:54 crc kubenswrapper[4912]: I1203 00:43:54.669759 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-4nzrg" Dec 03 00:43:55 crc kubenswrapper[4912]: I1203 00:43:55.264600 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-8c4wr" Dec 03 00:43:55 crc kubenswrapper[4912]: I1203 00:43:55.448094 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-4nzrg" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.684393 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7"] Dec 03 00:43:56 crc kubenswrapper[4912]: E1203 00:43:56.685254 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594281e0-91bf-4ca1-8937-b379f396d46d" containerName="registry-server" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.685269 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="594281e0-91bf-4ca1-8937-b379f396d46d" containerName="registry-server" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.685390 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="594281e0-91bf-4ca1-8937-b379f396d46d" containerName="registry-server" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.686557 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.690800 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-89dfp" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.696782 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7"] Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.770339 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-util\") pod \"e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.770412 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dhss\" (UniqueName: \"kubernetes.io/projected/f67e2276-c381-45a2-a5e8-c4a353fd8244-kube-api-access-9dhss\") pod \"e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.770460 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-bundle\") pod \"e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.872103 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-util\") pod \"e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.872196 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dhss\" (UniqueName: \"kubernetes.io/projected/f67e2276-c381-45a2-a5e8-c4a353fd8244-kube-api-access-9dhss\") pod \"e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.872361 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-bundle\") pod \"e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.872709 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-util\") pod \"e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.872832 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-bundle\") pod \"e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:56 crc kubenswrapper[4912]: I1203 00:43:56.893946 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dhss\" (UniqueName: \"kubernetes.io/projected/f67e2276-c381-45a2-a5e8-c4a353fd8244-kube-api-access-9dhss\") pod \"e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:57 crc kubenswrapper[4912]: I1203 00:43:57.016263 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:43:57 crc kubenswrapper[4912]: I1203 00:43:57.563802 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7"] Dec 03 00:43:57 crc kubenswrapper[4912]: W1203 00:43:57.581810 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf67e2276_c381_45a2_a5e8_c4a353fd8244.slice/crio-c55a1a2c9c298d36421cebcf407182b864a5c5b618f4e3554726c12083c06c97 WatchSource:0}: Error finding container c55a1a2c9c298d36421cebcf407182b864a5c5b618f4e3554726c12083c06c97: Status 404 returned error can't find the container with id c55a1a2c9c298d36421cebcf407182b864a5c5b618f4e3554726c12083c06c97 Dec 03 00:43:58 crc kubenswrapper[4912]: I1203 00:43:58.396273 4912 generic.go:334] "Generic (PLEG): container finished" podID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerID="f51208d51694d1573793f8b003f49c6b842be828a8f31858b60b766d0238fdb0" exitCode=0 Dec 03 00:43:58 crc kubenswrapper[4912]: I1203 00:43:58.396355 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" event={"ID":"f67e2276-c381-45a2-a5e8-c4a353fd8244","Type":"ContainerDied","Data":"f51208d51694d1573793f8b003f49c6b842be828a8f31858b60b766d0238fdb0"} Dec 03 00:43:58 crc kubenswrapper[4912]: I1203 00:43:58.396741 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" event={"ID":"f67e2276-c381-45a2-a5e8-c4a353fd8244","Type":"ContainerStarted","Data":"c55a1a2c9c298d36421cebcf407182b864a5c5b618f4e3554726c12083c06c97"} Dec 03 00:43:59 crc kubenswrapper[4912]: I1203 00:43:59.407002 4912 generic.go:334] "Generic (PLEG): container finished" podID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerID="46d168488be9c402992aa7933bb47ed2b1d5ac3e6dd4f4efeba40e1b569a69d1" exitCode=0 Dec 03 00:43:59 crc kubenswrapper[4912]: I1203 00:43:59.407067 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" event={"ID":"f67e2276-c381-45a2-a5e8-c4a353fd8244","Type":"ContainerDied","Data":"46d168488be9c402992aa7933bb47ed2b1d5ac3e6dd4f4efeba40e1b569a69d1"} Dec 03 00:44:00 crc kubenswrapper[4912]: I1203 00:44:00.436197 4912 generic.go:334] "Generic (PLEG): container finished" podID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerID="7e1424074f95569da84a1371e57fe6e365e648725e254429c2e002df28b574b2" exitCode=0 Dec 03 00:44:00 crc kubenswrapper[4912]: I1203 00:44:00.436243 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" event={"ID":"f67e2276-c381-45a2-a5e8-c4a353fd8244","Type":"ContainerDied","Data":"7e1424074f95569da84a1371e57fe6e365e648725e254429c2e002df28b574b2"} Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.800035 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.858350 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dhss\" (UniqueName: \"kubernetes.io/projected/f67e2276-c381-45a2-a5e8-c4a353fd8244-kube-api-access-9dhss\") pod \"f67e2276-c381-45a2-a5e8-c4a353fd8244\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.858537 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-bundle\") pod \"f67e2276-c381-45a2-a5e8-c4a353fd8244\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.858742 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-util\") pod \"f67e2276-c381-45a2-a5e8-c4a353fd8244\" (UID: \"f67e2276-c381-45a2-a5e8-c4a353fd8244\") " Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.859682 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-bundle" (OuterVolumeSpecName: "bundle") pod "f67e2276-c381-45a2-a5e8-c4a353fd8244" (UID: "f67e2276-c381-45a2-a5e8-c4a353fd8244"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.869447 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f67e2276-c381-45a2-a5e8-c4a353fd8244-kube-api-access-9dhss" (OuterVolumeSpecName: "kube-api-access-9dhss") pod "f67e2276-c381-45a2-a5e8-c4a353fd8244" (UID: "f67e2276-c381-45a2-a5e8-c4a353fd8244"). InnerVolumeSpecName "kube-api-access-9dhss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.873036 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-util" (OuterVolumeSpecName: "util") pod "f67e2276-c381-45a2-a5e8-c4a353fd8244" (UID: "f67e2276-c381-45a2-a5e8-c4a353fd8244"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.983370 4912 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-util\") on node \"crc\" DevicePath \"\"" Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.983480 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dhss\" (UniqueName: \"kubernetes.io/projected/f67e2276-c381-45a2-a5e8-c4a353fd8244-kube-api-access-9dhss\") on node \"crc\" DevicePath \"\"" Dec 03 00:44:01 crc kubenswrapper[4912]: I1203 00:44:01.983506 4912 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f67e2276-c381-45a2-a5e8-c4a353fd8244-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:44:02 crc kubenswrapper[4912]: I1203 00:44:02.462565 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" event={"ID":"f67e2276-c381-45a2-a5e8-c4a353fd8244","Type":"ContainerDied","Data":"c55a1a2c9c298d36421cebcf407182b864a5c5b618f4e3554726c12083c06c97"} Dec 03 00:44:02 crc kubenswrapper[4912]: I1203 00:44:02.462628 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c55a1a2c9c298d36421cebcf407182b864a5c5b618f4e3554726c12083c06c97" Dec 03 00:44:02 crc kubenswrapper[4912]: I1203 00:44:02.462721 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7" Dec 03 00:44:08 crc kubenswrapper[4912]: I1203 00:44:08.953231 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-c796f8559-s567r"] Dec 03 00:44:08 crc kubenswrapper[4912]: E1203 00:44:08.954179 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerName="extract" Dec 03 00:44:08 crc kubenswrapper[4912]: I1203 00:44:08.954195 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerName="extract" Dec 03 00:44:08 crc kubenswrapper[4912]: E1203 00:44:08.954210 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerName="pull" Dec 03 00:44:08 crc kubenswrapper[4912]: I1203 00:44:08.954218 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerName="pull" Dec 03 00:44:08 crc kubenswrapper[4912]: E1203 00:44:08.954248 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerName="util" Dec 03 00:44:08 crc kubenswrapper[4912]: I1203 00:44:08.954255 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerName="util" Dec 03 00:44:08 crc kubenswrapper[4912]: I1203 00:44:08.954409 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f67e2276-c381-45a2-a5e8-c4a353fd8244" containerName="extract" Dec 03 00:44:08 crc kubenswrapper[4912]: I1203 00:44:08.955132 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" Dec 03 00:44:08 crc kubenswrapper[4912]: I1203 00:44:08.958872 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-jrnd6" Dec 03 00:44:08 crc kubenswrapper[4912]: I1203 00:44:08.980719 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-c796f8559-s567r"] Dec 03 00:44:09 crc kubenswrapper[4912]: I1203 00:44:09.012937 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kwdh\" (UniqueName: \"kubernetes.io/projected/e96cfd80-56de-4ab0-98b0-8bdcb324e498-kube-api-access-8kwdh\") pod \"openstack-operator-controller-operator-c796f8559-s567r\" (UID: \"e96cfd80-56de-4ab0-98b0-8bdcb324e498\") " pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" Dec 03 00:44:09 crc kubenswrapper[4912]: I1203 00:44:09.114441 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kwdh\" (UniqueName: \"kubernetes.io/projected/e96cfd80-56de-4ab0-98b0-8bdcb324e498-kube-api-access-8kwdh\") pod \"openstack-operator-controller-operator-c796f8559-s567r\" (UID: \"e96cfd80-56de-4ab0-98b0-8bdcb324e498\") " pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" Dec 03 00:44:09 crc kubenswrapper[4912]: I1203 00:44:09.135799 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kwdh\" (UniqueName: \"kubernetes.io/projected/e96cfd80-56de-4ab0-98b0-8bdcb324e498-kube-api-access-8kwdh\") pod \"openstack-operator-controller-operator-c796f8559-s567r\" (UID: \"e96cfd80-56de-4ab0-98b0-8bdcb324e498\") " pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" Dec 03 00:44:09 crc kubenswrapper[4912]: I1203 00:44:09.279224 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" Dec 03 00:44:09 crc kubenswrapper[4912]: I1203 00:44:09.790237 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-c796f8559-s567r"] Dec 03 00:44:10 crc kubenswrapper[4912]: I1203 00:44:10.553676 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" event={"ID":"e96cfd80-56de-4ab0-98b0-8bdcb324e498","Type":"ContainerStarted","Data":"5dff4d5dc1dbdef30dbc1b287e9fb6b8605683dd9fa1f830e4240f51b65d500d"} Dec 03 00:44:15 crc kubenswrapper[4912]: I1203 00:44:15.604694 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" event={"ID":"e96cfd80-56de-4ab0-98b0-8bdcb324e498","Type":"ContainerStarted","Data":"53deca348b7444a00ee7776ec8c3e52db73af278a148f8225edc70653235ed75"} Dec 03 00:44:15 crc kubenswrapper[4912]: I1203 00:44:15.605699 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" Dec 03 00:44:15 crc kubenswrapper[4912]: I1203 00:44:15.645342 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" podStartSLOduration=2.923765697 podStartE2EDuration="7.645306234s" podCreationTimestamp="2025-12-03 00:44:08 +0000 UTC" firstStartedPulling="2025-12-03 00:44:09.794278092 +0000 UTC m=+1235.436298652" lastFinishedPulling="2025-12-03 00:44:14.515818629 +0000 UTC m=+1240.157839189" observedRunningTime="2025-12-03 00:44:15.639728162 +0000 UTC m=+1241.281748802" watchObservedRunningTime="2025-12-03 00:44:15.645306234 +0000 UTC m=+1241.287326804" Dec 03 00:44:19 crc kubenswrapper[4912]: I1203 00:44:19.283944 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-c796f8559-s567r" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.675540 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.677943 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.680195 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-n4qfm" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.701002 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.702774 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.708182 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.709470 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.714417 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-2lt2j" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.714454 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-qvmpn" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.716655 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.723374 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.730601 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.744078 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.745597 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.750544 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-5rqgz" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.761680 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.818670 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s57k2\" (UniqueName: \"kubernetes.io/projected/5118703e-2fdb-41d6-a76e-ab5d3f6adb95-kube-api-access-s57k2\") pod \"designate-operator-controller-manager-78b4bc895b-tqphq\" (UID: \"5118703e-2fdb-41d6-a76e-ab5d3f6adb95\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.818940 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpltn\" (UniqueName: \"kubernetes.io/projected/70a932ec-65ac-4616-870b-9297bd5d0a28-kube-api-access-rpltn\") pod \"cinder-operator-controller-manager-859b6ccc6-ds5mt\" (UID: \"70a932ec-65ac-4616-870b-9297bd5d0a28\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.819350 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vczzz\" (UniqueName: \"kubernetes.io/projected/56cdea32-a9a4-410b-9a9c-c237746b0582-kube-api-access-vczzz\") pod \"barbican-operator-controller-manager-7d9dfd778-twd26\" (UID: \"56cdea32-a9a4-410b-9a9c-c237746b0582\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.859696 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.863351 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.867649 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-qb5kl" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.875924 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.878122 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.885330 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.888828 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.891015 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-lbpf2" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.899591 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-lzff2"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.901274 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.914214 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-whj5c" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.914396 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.924304 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vczzz\" (UniqueName: \"kubernetes.io/projected/56cdea32-a9a4-410b-9a9c-c237746b0582-kube-api-access-vczzz\") pod \"barbican-operator-controller-manager-7d9dfd778-twd26\" (UID: \"56cdea32-a9a4-410b-9a9c-c237746b0582\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.924383 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtt7v\" (UniqueName: \"kubernetes.io/projected/045ab366-9392-40f9-94f5-95c983f65176-kube-api-access-gtt7v\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.924410 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g6fg\" (UniqueName: \"kubernetes.io/projected/d2643e7f-8db2-4374-b0d9-05a2130795aa-kube-api-access-8g6fg\") pod \"glance-operator-controller-manager-77987cd8cd-g4n8g\" (UID: \"d2643e7f-8db2-4374-b0d9-05a2130795aa\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.924446 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.924466 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhc5w\" (UniqueName: \"kubernetes.io/projected/27605b80-0a7f-4b6b-9729-d0b6eaa0a74b-kube-api-access-rhc5w\") pod \"horizon-operator-controller-manager-68c6d99b8f-rb4z7\" (UID: \"27605b80-0a7f-4b6b-9729-d0b6eaa0a74b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.924488 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s57k2\" (UniqueName: \"kubernetes.io/projected/5118703e-2fdb-41d6-a76e-ab5d3f6adb95-kube-api-access-s57k2\") pod \"designate-operator-controller-manager-78b4bc895b-tqphq\" (UID: \"5118703e-2fdb-41d6-a76e-ab5d3f6adb95\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.924523 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpltn\" (UniqueName: \"kubernetes.io/projected/70a932ec-65ac-4616-870b-9297bd5d0a28-kube-api-access-rpltn\") pod \"cinder-operator-controller-manager-859b6ccc6-ds5mt\" (UID: \"70a932ec-65ac-4616-870b-9297bd5d0a28\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.924808 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm8cc\" (UniqueName: \"kubernetes.io/projected/1df68f98-9a63-43ca-8e47-5c34c1e52581-kube-api-access-dm8cc\") pod \"heat-operator-controller-manager-5f64f6f8bb-jskq6\" (UID: \"1df68f98-9a63-43ca-8e47-5c34c1e52581\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.933603 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.935114 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.944026 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-lzff2"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.951884 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-7hc7w" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.955100 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.970558 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.972418 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.974912 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-w85x8" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.978024 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vczzz\" (UniqueName: \"kubernetes.io/projected/56cdea32-a9a4-410b-9a9c-c237746b0582-kube-api-access-vczzz\") pod \"barbican-operator-controller-manager-7d9dfd778-twd26\" (UID: \"56cdea32-a9a4-410b-9a9c-c237746b0582\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.981216 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpltn\" (UniqueName: \"kubernetes.io/projected/70a932ec-65ac-4616-870b-9297bd5d0a28-kube-api-access-rpltn\") pod \"cinder-operator-controller-manager-859b6ccc6-ds5mt\" (UID: \"70a932ec-65ac-4616-870b-9297bd5d0a28\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.990129 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s57k2\" (UniqueName: \"kubernetes.io/projected/5118703e-2fdb-41d6-a76e-ab5d3f6adb95-kube-api-access-s57k2\") pod \"designate-operator-controller-manager-78b4bc895b-tqphq\" (UID: \"5118703e-2fdb-41d6-a76e-ab5d3f6adb95\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.990238 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2"] Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.991834 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" Dec 03 00:44:39 crc kubenswrapper[4912]: I1203 00:44:39.999512 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.001558 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.006389 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.007347 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-t8x4d" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.027692 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtt7v\" (UniqueName: \"kubernetes.io/projected/045ab366-9392-40f9-94f5-95c983f65176-kube-api-access-gtt7v\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.027775 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g6fg\" (UniqueName: \"kubernetes.io/projected/d2643e7f-8db2-4374-b0d9-05a2130795aa-kube-api-access-8g6fg\") pod \"glance-operator-controller-manager-77987cd8cd-g4n8g\" (UID: \"d2643e7f-8db2-4374-b0d9-05a2130795aa\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.027814 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.027838 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhc5w\" (UniqueName: \"kubernetes.io/projected/27605b80-0a7f-4b6b-9729-d0b6eaa0a74b-kube-api-access-rhc5w\") pod \"horizon-operator-controller-manager-68c6d99b8f-rb4z7\" (UID: \"27605b80-0a7f-4b6b-9729-d0b6eaa0a74b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.027920 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm8cc\" (UniqueName: \"kubernetes.io/projected/1df68f98-9a63-43ca-8e47-5c34c1e52581-kube-api-access-dm8cc\") pod \"heat-operator-controller-manager-5f64f6f8bb-jskq6\" (UID: \"1df68f98-9a63-43ca-8e47-5c34c1e52581\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" Dec 03 00:44:40 crc kubenswrapper[4912]: E1203 00:44:40.028944 4912 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:40 crc kubenswrapper[4912]: E1203 00:44:40.029015 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert podName:045ab366-9392-40f9-94f5-95c983f65176 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:40.528989972 +0000 UTC m=+1266.171010532 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert") pod "infra-operator-controller-manager-57548d458d-lzff2" (UID: "045ab366-9392-40f9-94f5-95c983f65176") : secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.029444 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.039354 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.050254 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.057329 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.058391 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.058689 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.076042 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-jm2pj" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.076725 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-jwzlz" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.085479 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g6fg\" (UniqueName: \"kubernetes.io/projected/d2643e7f-8db2-4374-b0d9-05a2130795aa-kube-api-access-8g6fg\") pod \"glance-operator-controller-manager-77987cd8cd-g4n8g\" (UID: \"d2643e7f-8db2-4374-b0d9-05a2130795aa\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.088991 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm8cc\" (UniqueName: \"kubernetes.io/projected/1df68f98-9a63-43ca-8e47-5c34c1e52581-kube-api-access-dm8cc\") pod \"heat-operator-controller-manager-5f64f6f8bb-jskq6\" (UID: \"1df68f98-9a63-43ca-8e47-5c34c1e52581\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.089589 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtt7v\" (UniqueName: \"kubernetes.io/projected/045ab366-9392-40f9-94f5-95c983f65176-kube-api-access-gtt7v\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.099953 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhc5w\" (UniqueName: \"kubernetes.io/projected/27605b80-0a7f-4b6b-9729-d0b6eaa0a74b-kube-api-access-rhc5w\") pod \"horizon-operator-controller-manager-68c6d99b8f-rb4z7\" (UID: \"27605b80-0a7f-4b6b-9729-d0b6eaa0a74b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.102484 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.122517 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.133917 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l7mh\" (UniqueName: \"kubernetes.io/projected/9b23d447-fdc8-47a1-af38-a569b6ad2a8f-kube-api-access-8l7mh\") pod \"keystone-operator-controller-manager-7765d96ddf-txtqb\" (UID: \"9b23d447-fdc8-47a1-af38-a569b6ad2a8f\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.134045 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc7qv\" (UniqueName: \"kubernetes.io/projected/8ddf2fc1-446c-4a26-987a-f67ec4c51455-kube-api-access-hc7qv\") pod \"manila-operator-controller-manager-7c79b5df47-fxhh2\" (UID: \"8ddf2fc1-446c-4a26-987a-f67ec4c51455\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.134094 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jg95\" (UniqueName: \"kubernetes.io/projected/7b515331-ceab-4f45-9880-719c72dfcc4c-kube-api-access-7jg95\") pod \"ironic-operator-controller-manager-6c548fd776-626ss\" (UID: \"7b515331-ceab-4f45-9880-719c72dfcc4c\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.173714 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.177004 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.188572 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-6b4xt" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.196862 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.231549 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.237893 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.238364 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l7mh\" (UniqueName: \"kubernetes.io/projected/9b23d447-fdc8-47a1-af38-a569b6ad2a8f-kube-api-access-8l7mh\") pod \"keystone-operator-controller-manager-7765d96ddf-txtqb\" (UID: \"9b23d447-fdc8-47a1-af38-a569b6ad2a8f\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.238481 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc7qv\" (UniqueName: \"kubernetes.io/projected/8ddf2fc1-446c-4a26-987a-f67ec4c51455-kube-api-access-hc7qv\") pod \"manila-operator-controller-manager-7c79b5df47-fxhh2\" (UID: \"8ddf2fc1-446c-4a26-987a-f67ec4c51455\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.238534 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6cn2\" (UniqueName: \"kubernetes.io/projected/a7db7741-5afc-46bc-84c4-71c233ca9dde-kube-api-access-w6cn2\") pod \"mariadb-operator-controller-manager-56bbcc9d85-rjqjr\" (UID: \"a7db7741-5afc-46bc-84c4-71c233ca9dde\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.238573 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jg95\" (UniqueName: \"kubernetes.io/projected/7b515331-ceab-4f45-9880-719c72dfcc4c-kube-api-access-7jg95\") pod \"ironic-operator-controller-manager-6c548fd776-626ss\" (UID: \"7b515331-ceab-4f45-9880-719c72dfcc4c\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.238599 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fts8l\" (UniqueName: \"kubernetes.io/projected/6848c014-0383-482f-9d46-232cc8fafbe1-kube-api-access-fts8l\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-mjphw\" (UID: \"6848c014-0383-482f-9d46-232cc8fafbe1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.244356 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.246088 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.257375 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-pt4br" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.273563 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc7qv\" (UniqueName: \"kubernetes.io/projected/8ddf2fc1-446c-4a26-987a-f67ec4c51455-kube-api-access-hc7qv\") pod \"manila-operator-controller-manager-7c79b5df47-fxhh2\" (UID: \"8ddf2fc1-446c-4a26-987a-f67ec4c51455\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.289248 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.291773 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l7mh\" (UniqueName: \"kubernetes.io/projected/9b23d447-fdc8-47a1-af38-a569b6ad2a8f-kube-api-access-8l7mh\") pod \"keystone-operator-controller-manager-7765d96ddf-txtqb\" (UID: \"9b23d447-fdc8-47a1-af38-a569b6ad2a8f\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.294087 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-5n67v"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.304798 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.318341 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-pz8xv" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.340290 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6cn2\" (UniqueName: \"kubernetes.io/projected/a7db7741-5afc-46bc-84c4-71c233ca9dde-kube-api-access-w6cn2\") pod \"mariadb-operator-controller-manager-56bbcc9d85-rjqjr\" (UID: \"a7db7741-5afc-46bc-84c4-71c233ca9dde\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.340374 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fts8l\" (UniqueName: \"kubernetes.io/projected/6848c014-0383-482f-9d46-232cc8fafbe1-kube-api-access-fts8l\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-mjphw\" (UID: \"6848c014-0383-482f-9d46-232cc8fafbe1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.340437 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltcqf\" (UniqueName: \"kubernetes.io/projected/baaa7298-1635-4de5-af87-68db95102082-kube-api-access-ltcqf\") pod \"octavia-operator-controller-manager-998648c74-dtgrm\" (UID: \"baaa7298-1635-4de5-af87-68db95102082\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.340499 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c65n\" (UniqueName: \"kubernetes.io/projected/de517f42-de01-4dde-ac22-1a651a38926b-kube-api-access-2c65n\") pod \"nova-operator-controller-manager-697bc559fc-bdgdj\" (UID: \"de517f42-de01-4dde-ac22-1a651a38926b\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.348488 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.357180 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jg95\" (UniqueName: \"kubernetes.io/projected/7b515331-ceab-4f45-9880-719c72dfcc4c-kube-api-access-7jg95\") pod \"ironic-operator-controller-manager-6c548fd776-626ss\" (UID: \"7b515331-ceab-4f45-9880-719c72dfcc4c\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.362654 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.462544 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9svhq\" (UniqueName: \"kubernetes.io/projected/7ec29496-d0ac-42a2-acea-7037dfde41fb-kube-api-access-9svhq\") pod \"placement-operator-controller-manager-78f8948974-5n67v\" (UID: \"7ec29496-d0ac-42a2-acea-7037dfde41fb\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.463093 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltcqf\" (UniqueName: \"kubernetes.io/projected/baaa7298-1635-4de5-af87-68db95102082-kube-api-access-ltcqf\") pod \"octavia-operator-controller-manager-998648c74-dtgrm\" (UID: \"baaa7298-1635-4de5-af87-68db95102082\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.463212 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c65n\" (UniqueName: \"kubernetes.io/projected/de517f42-de01-4dde-ac22-1a651a38926b-kube-api-access-2c65n\") pod \"nova-operator-controller-manager-697bc559fc-bdgdj\" (UID: \"de517f42-de01-4dde-ac22-1a651a38926b\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.464033 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.472985 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.526916 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.557857 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-wmlhw" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.558056 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.568354 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p92vq\" (UniqueName: \"kubernetes.io/projected/fd3cb745-7c4b-484b-93a5-7638dcd4d738-kube-api-access-p92vq\") pod \"ovn-operator-controller-manager-b6456fdb6-vw7dk\" (UID: \"fd3cb745-7c4b-484b-93a5-7638dcd4d738\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.568471 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.568568 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9svhq\" (UniqueName: \"kubernetes.io/projected/7ec29496-d0ac-42a2-acea-7037dfde41fb-kube-api-access-9svhq\") pod \"placement-operator-controller-manager-78f8948974-5n67v\" (UID: \"7ec29496-d0ac-42a2-acea-7037dfde41fb\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" Dec 03 00:44:40 crc kubenswrapper[4912]: E1203 00:44:40.569187 4912 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:40 crc kubenswrapper[4912]: E1203 00:44:40.569245 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert podName:045ab366-9392-40f9-94f5-95c983f65176 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:41.569230249 +0000 UTC m=+1267.211250809 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert") pod "infra-operator-controller-manager-57548d458d-lzff2" (UID: "045ab366-9392-40f9-94f5-95c983f65176") : secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.572389 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c65n\" (UniqueName: \"kubernetes.io/projected/de517f42-de01-4dde-ac22-1a651a38926b-kube-api-access-2c65n\") pod \"nova-operator-controller-manager-697bc559fc-bdgdj\" (UID: \"de517f42-de01-4dde-ac22-1a651a38926b\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.572920 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.573671 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fts8l\" (UniqueName: \"kubernetes.io/projected/6848c014-0383-482f-9d46-232cc8fafbe1-kube-api-access-fts8l\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-mjphw\" (UID: \"6848c014-0383-482f-9d46-232cc8fafbe1\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.574275 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.574368 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.577972 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-gffxw" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.596732 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6cn2\" (UniqueName: \"kubernetes.io/projected/a7db7741-5afc-46bc-84c4-71c233ca9dde-kube-api-access-w6cn2\") pod \"mariadb-operator-controller-manager-56bbcc9d85-rjqjr\" (UID: \"a7db7741-5afc-46bc-84c4-71c233ca9dde\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.608281 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltcqf\" (UniqueName: \"kubernetes.io/projected/baaa7298-1635-4de5-af87-68db95102082-kube-api-access-ltcqf\") pod \"octavia-operator-controller-manager-998648c74-dtgrm\" (UID: \"baaa7298-1635-4de5-af87-68db95102082\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.633179 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.633481 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9svhq\" (UniqueName: \"kubernetes.io/projected/7ec29496-d0ac-42a2-acea-7037dfde41fb-kube-api-access-9svhq\") pod \"placement-operator-controller-manager-78f8948974-5n67v\" (UID: \"7ec29496-d0ac-42a2-acea-7037dfde41fb\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.639688 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-5n67v"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.655200 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.664915 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.667074 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.669868 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-mcb6v" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.670757 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hfgk\" (UniqueName: \"kubernetes.io/projected/3e85fc81-3fbd-4af2-8c98-89892a4f3579-kube-api-access-9hfgk\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.670895 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.670966 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p92vq\" (UniqueName: \"kubernetes.io/projected/fd3cb745-7c4b-484b-93a5-7638dcd4d738-kube-api-access-p92vq\") pod \"ovn-operator-controller-manager-b6456fdb6-vw7dk\" (UID: \"fd3cb745-7c4b-484b-93a5-7638dcd4d738\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.683557 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.685819 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.686466 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.692932 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-c5gqc" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.713004 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.719518 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.744066 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.745706 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p92vq\" (UniqueName: \"kubernetes.io/projected/fd3cb745-7c4b-484b-93a5-7638dcd4d738-kube-api-access-p92vq\") pod \"ovn-operator-controller-manager-b6456fdb6-vw7dk\" (UID: \"fd3cb745-7c4b-484b-93a5-7638dcd4d738\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.746174 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.772180 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hfgk\" (UniqueName: \"kubernetes.io/projected/3e85fc81-3fbd-4af2-8c98-89892a4f3579-kube-api-access-9hfgk\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.772267 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfnmj\" (UniqueName: \"kubernetes.io/projected/82b74254-de20-48a4-bfae-21afed7954cd-kube-api-access-kfnmj\") pod \"telemetry-operator-controller-manager-5d66d5c957-xgv9t\" (UID: \"82b74254-de20-48a4-bfae-21afed7954cd\") " pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.772539 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:40 crc kubenswrapper[4912]: E1203 00:44:40.773072 4912 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:40 crc kubenswrapper[4912]: E1203 00:44:40.773126 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert podName:3e85fc81-3fbd-4af2-8c98-89892a4f3579 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:41.273108804 +0000 UTC m=+1266.915129364 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" (UID: "3e85fc81-3fbd-4af2-8c98-89892a4f3579") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.792007 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.816022 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.822942 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-7rstc"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.826865 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hfgk\" (UniqueName: \"kubernetes.io/projected/3e85fc81-3fbd-4af2-8c98-89892a4f3579-kube-api-access-9hfgk\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.828165 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.842391 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-rqdc9" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.844759 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-7rstc"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.862902 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.874041 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwxxp\" (UniqueName: \"kubernetes.io/projected/2ab18e2a-3358-44e6-ac55-80b072619e73-kube-api-access-vwxxp\") pod \"swift-operator-controller-manager-5f8c65bbfc-8x84c\" (UID: \"2ab18e2a-3358-44e6-ac55-80b072619e73\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.874126 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfnmj\" (UniqueName: \"kubernetes.io/projected/82b74254-de20-48a4-bfae-21afed7954cd-kube-api-access-kfnmj\") pod \"telemetry-operator-controller-manager-5d66d5c957-xgv9t\" (UID: \"82b74254-de20-48a4-bfae-21afed7954cd\") " pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.896364 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.901973 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.917950 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7"] Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.936328 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-tqnpt" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.937198 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfnmj\" (UniqueName: \"kubernetes.io/projected/82b74254-de20-48a4-bfae-21afed7954cd-kube-api-access-kfnmj\") pod \"telemetry-operator-controller-manager-5d66d5c957-xgv9t\" (UID: \"82b74254-de20-48a4-bfae-21afed7954cd\") " pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.978722 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9xx9\" (UniqueName: \"kubernetes.io/projected/b5544f39-6c3a-4737-a8e3-389a6e66ba40-kube-api-access-r9xx9\") pod \"test-operator-controller-manager-5854674fcc-7rstc\" (UID: \"b5544f39-6c3a-4737-a8e3-389a6e66ba40\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" Dec 03 00:44:40 crc kubenswrapper[4912]: I1203 00:44:40.978861 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwxxp\" (UniqueName: \"kubernetes.io/projected/2ab18e2a-3358-44e6-ac55-80b072619e73-kube-api-access-vwxxp\") pod \"swift-operator-controller-manager-5f8c65bbfc-8x84c\" (UID: \"2ab18e2a-3358-44e6-ac55-80b072619e73\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.000735 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86464467f8-sq576"] Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.002544 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.006888 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.007272 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-s222x" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.014282 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwxxp\" (UniqueName: \"kubernetes.io/projected/2ab18e2a-3358-44e6-ac55-80b072619e73-kube-api-access-vwxxp\") pod \"swift-operator-controller-manager-5f8c65bbfc-8x84c\" (UID: \"2ab18e2a-3358-44e6-ac55-80b072619e73\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.020468 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.061398 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86464467f8-sq576"] Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.080695 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqtpv\" (UniqueName: \"kubernetes.io/projected/ea0ec30d-2c13-49fd-883b-767024ff632c-kube-api-access-nqtpv\") pod \"watcher-operator-controller-manager-769dc69bc-jqzw7\" (UID: \"ea0ec30d-2c13-49fd-883b-767024ff632c\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.080747 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9xx9\" (UniqueName: \"kubernetes.io/projected/b5544f39-6c3a-4737-a8e3-389a6e66ba40-kube-api-access-r9xx9\") pod \"test-operator-controller-manager-5854674fcc-7rstc\" (UID: \"b5544f39-6c3a-4737-a8e3-389a6e66ba40\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.107021 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt"] Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.109684 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.112009 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-x6rx7" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.126492 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9xx9\" (UniqueName: \"kubernetes.io/projected/b5544f39-6c3a-4737-a8e3-389a6e66ba40-kube-api-access-r9xx9\") pod \"test-operator-controller-manager-5854674fcc-7rstc\" (UID: \"b5544f39-6c3a-4737-a8e3-389a6e66ba40\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.163355 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt"] Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.182250 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.182312 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nrwz\" (UniqueName: \"kubernetes.io/projected/10227b7f-ff98-4f4a-862f-3ef78b4e7737-kube-api-access-5nrwz\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.182479 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.182910 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqtpv\" (UniqueName: \"kubernetes.io/projected/ea0ec30d-2c13-49fd-883b-767024ff632c-kube-api-access-nqtpv\") pod \"watcher-operator-controller-manager-769dc69bc-jqzw7\" (UID: \"ea0ec30d-2c13-49fd-883b-767024ff632c\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.203568 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.217495 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqtpv\" (UniqueName: \"kubernetes.io/projected/ea0ec30d-2c13-49fd-883b-767024ff632c-kube-api-access-nqtpv\") pod \"watcher-operator-controller-manager-769dc69bc-jqzw7\" (UID: \"ea0ec30d-2c13-49fd-883b-767024ff632c\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.245662 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.250541 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.284246 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.284278 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nrwz\" (UniqueName: \"kubernetes.io/projected/10227b7f-ff98-4f4a-862f-3ef78b4e7737-kube-api-access-5nrwz\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.284316 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.284340 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.284381 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8ccx\" (UniqueName: \"kubernetes.io/projected/b54a05e7-9cfd-401d-8e5d-87c8ed5ad321-kube-api-access-f8ccx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-llhvt\" (UID: \"b54a05e7-9cfd-401d-8e5d-87c8ed5ad321\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.284592 4912 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.284654 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:41.784618081 +0000 UTC m=+1267.426638641 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "metrics-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.285089 4912 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.285118 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:41.785110074 +0000 UTC m=+1267.427130634 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "webhook-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.286761 4912 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.286813 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert podName:3e85fc81-3fbd-4af2-8c98-89892a4f3579 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:42.286800041 +0000 UTC m=+1267.928820601 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" (UID: "3e85fc81-3fbd-4af2-8c98-89892a4f3579") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.296735 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.318424 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nrwz\" (UniqueName: \"kubernetes.io/projected/10227b7f-ff98-4f4a-862f-3ef78b4e7737-kube-api-access-5nrwz\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.405146 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8ccx\" (UniqueName: \"kubernetes.io/projected/b54a05e7-9cfd-401d-8e5d-87c8ed5ad321-kube-api-access-f8ccx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-llhvt\" (UID: \"b54a05e7-9cfd-401d-8e5d-87c8ed5ad321\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.435695 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26"] Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.453597 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8ccx\" (UniqueName: \"kubernetes.io/projected/b54a05e7-9cfd-401d-8e5d-87c8ed5ad321-kube-api-access-f8ccx\") pod \"rabbitmq-cluster-operator-manager-668c99d594-llhvt\" (UID: \"b54a05e7-9cfd-401d-8e5d-87c8ed5ad321\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.610614 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.610822 4912 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.610887 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert podName:045ab366-9392-40f9-94f5-95c983f65176 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:43.610866909 +0000 UTC m=+1269.252887469 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert") pod "infra-operator-controller-manager-57548d458d-lzff2" (UID: "045ab366-9392-40f9-94f5-95c983f65176") : secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.729498 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.735595 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt"] Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.816662 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.816959 4912 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.817013 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.817116 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:42.817087217 +0000 UTC m=+1268.459107777 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "metrics-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.817316 4912 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: E1203 00:44:41.817480 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:42.817460498 +0000 UTC m=+1268.459481058 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "webhook-server-cert" not found Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.890871 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" event={"ID":"70a932ec-65ac-4616-870b-9297bd5d0a28","Type":"ContainerStarted","Data":"755e4faf9d57d5545897dfcb0085d4996cf71c5c9305b59d69995ec45032d19e"} Dec 03 00:44:41 crc kubenswrapper[4912]: I1203 00:44:41.893073 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" event={"ID":"56cdea32-a9a4-410b-9a9c-c237746b0582","Type":"ContainerStarted","Data":"6c1836f0516253f5df3b04c6d4cdef293e18f2536e70bf4c51f31e6b610d7a5d"} Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.330750 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.330961 4912 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.331297 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert podName:3e85fc81-3fbd-4af2-8c98-89892a4f3579 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:44.331272888 +0000 UTC m=+1269.973293448 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" (UID: "3e85fc81-3fbd-4af2-8c98-89892a4f3579") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.386920 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.428662 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-5n67v"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.451883 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.506017 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj"] Dec 03 00:44:42 crc kubenswrapper[4912]: W1203 00:44:42.517629 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd3cb745_7c4b_484b_93a5_7638dcd4d738.slice/crio-d676958352a72dcc02049875a6d2035305f9e16ae83a80e8127dc9c8f1d241b2 WatchSource:0}: Error finding container d676958352a72dcc02049875a6d2035305f9e16ae83a80e8127dc9c8f1d241b2: Status 404 returned error can't find the container with id d676958352a72dcc02049875a6d2035305f9e16ae83a80e8127dc9c8f1d241b2 Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.524338 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.532472 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk"] Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.534636 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p92vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-vw7dk_openstack-operators(fd3cb745-7c4b-484b-93a5-7638dcd4d738): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.534731 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fts8l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-mjphw_openstack-operators(6848c014-0383-482f-9d46-232cc8fafbe1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.541766 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6"] Dec 03 00:44:42 crc kubenswrapper[4912]: W1203 00:44:42.543968 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ab18e2a_3358_44e6_ac55_80b072619e73.slice/crio-e6e03d7d84fd67c3c296e0596bb1e4697080d16f5d648c1d696b69602ecd9180 WatchSource:0}: Error finding container e6e03d7d84fd67c3c296e0596bb1e4697080d16f5d648c1d696b69602ecd9180: Status 404 returned error can't find the container with id e6e03d7d84fd67c3c296e0596bb1e4697080d16f5d648c1d696b69602ecd9180 Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.544155 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p92vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-vw7dk_openstack-operators(fd3cb745-7c4b-484b-93a5-7638dcd4d738): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.544192 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fts8l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-mjphw_openstack-operators(6848c014-0383-482f-9d46-232cc8fafbe1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.545531 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" podUID="6848c014-0383-482f-9d46-232cc8fafbe1" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.545561 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" podUID="fd3cb745-7c4b-484b-93a5-7638dcd4d738" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.552738 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vwxxp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-8x84c_openstack-operators(2ab18e2a-3358-44e6-ac55-80b072619e73): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.554936 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7"] Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.559641 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w6cn2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-rjqjr_openstack-operators(a7db7741-5afc-46bc-84c4-71c233ca9dde): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.559800 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vwxxp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-8x84c_openstack-operators(2ab18e2a-3358-44e6-ac55-80b072619e73): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.561690 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" podUID="2ab18e2a-3358-44e6-ac55-80b072619e73" Dec 03 00:44:42 crc kubenswrapper[4912]: W1203 00:44:42.563051 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82b74254_de20_48a4_bfae_21afed7954cd.slice/crio-3efe394473f8f79631fddbedbe94c62d9c08dbf8c2151fc7b328233c928cd554 WatchSource:0}: Error finding container 3efe394473f8f79631fddbedbe94c62d9c08dbf8c2151fc7b328233c928cd554: Status 404 returned error can't find the container with id 3efe394473f8f79631fddbedbe94c62d9c08dbf8c2151fc7b328233c928cd554 Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.563662 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w6cn2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-rjqjr_openstack-operators(a7db7741-5afc-46bc-84c4-71c233ca9dde): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.564743 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" podUID="a7db7741-5afc-46bc-84c4-71c233ca9dde" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.575029 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.150:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kfnmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5d66d5c957-xgv9t_openstack-operators(82b74254-de20-48a4-bfae-21afed7954cd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.578507 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kfnmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5d66d5c957-xgv9t_openstack-operators(82b74254-de20-48a4-bfae-21afed7954cd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.579726 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" podUID="82b74254-de20-48a4-bfae-21afed7954cd" Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.598310 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.605492 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.614586 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.622378 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.630693 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.639765 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.649034 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t"] Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.663927 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7"] Dec 03 00:44:42 crc kubenswrapper[4912]: W1203 00:44:42.705202 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb54a05e7_9cfd_401d_8e5d_87c8ed5ad321.slice/crio-7113ae583797635f4fab83353ec8d14a347a29a964c5bc3279f26084c5be5ff2 WatchSource:0}: Error finding container 7113ae583797635f4fab83353ec8d14a347a29a964c5bc3279f26084c5be5ff2: Status 404 returned error can't find the container with id 7113ae583797635f4fab83353ec8d14a347a29a964c5bc3279f26084c5be5ff2 Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.708516 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt"] Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.711286 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f8ccx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-llhvt_openstack-operators(b54a05e7-9cfd-401d-8e5d-87c8ed5ad321): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.712504 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" podUID="b54a05e7-9cfd-401d-8e5d-87c8ed5ad321" Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.717643 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-7rstc"] Dec 03 00:44:42 crc kubenswrapper[4912]: W1203 00:44:42.719895 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5544f39_6c3a_4737_a8e3_389a6e66ba40.slice/crio-0552def0934dc231253cc22b7d8add649e8b5759ad117ef5a77ce7a55640eadb WatchSource:0}: Error finding container 0552def0934dc231253cc22b7d8add649e8b5759ad117ef5a77ce7a55640eadb: Status 404 returned error can't find the container with id 0552def0934dc231253cc22b7d8add649e8b5759ad117ef5a77ce7a55640eadb Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.723873 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r9xx9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-7rstc_openstack-operators(b5544f39-6c3a-4737-a8e3-389a6e66ba40): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.728607 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r9xx9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-7rstc_openstack-operators(b5544f39-6c3a-4737-a8e3-389a6e66ba40): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.729697 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" podUID="b5544f39-6c3a-4737-a8e3-389a6e66ba40" Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.861538 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.861671 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.862098 4912 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.862168 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:44.86214872 +0000 UTC m=+1270.504169290 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "webhook-server-cert" not found Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.862752 4912 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.862805 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:44.862793527 +0000 UTC m=+1270.504814087 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "metrics-server-cert" not found Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.904685 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" event={"ID":"7ec29496-d0ac-42a2-acea-7037dfde41fb","Type":"ContainerStarted","Data":"a54a9be99b2e85253760a8d76077ebafe9f2e13b8285d8d9eccd349b274c5f28"} Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.907230 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" event={"ID":"a7db7741-5afc-46bc-84c4-71c233ca9dde","Type":"ContainerStarted","Data":"d4626e6912e1c47cdb418e2da969c530e6f0d88fdaf30aa96e4ef573522a9a80"} Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.909516 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" event={"ID":"7b515331-ceab-4f45-9880-719c72dfcc4c","Type":"ContainerStarted","Data":"61c6b4bac3a846fef56b2ee51a16cdfc63c83b5553169a158c0eb51b75c306e7"} Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.913716 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" podUID="a7db7741-5afc-46bc-84c4-71c233ca9dde" Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.913908 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" event={"ID":"1df68f98-9a63-43ca-8e47-5c34c1e52581","Type":"ContainerStarted","Data":"dbf84f91645bbc90add28dbaf1183d9eefa76dfe399c1c307fb544199f099d24"} Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.920809 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" event={"ID":"baaa7298-1635-4de5-af87-68db95102082","Type":"ContainerStarted","Data":"1f7179a6ec541fb311cdba43013d9b85c5833cbaf688ec37838bd3c1e03a45bf"} Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.934182 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" event={"ID":"27605b80-0a7f-4b6b-9729-d0b6eaa0a74b","Type":"ContainerStarted","Data":"2d6825fe760d36001c539dc4de1b84f9894252445b54cfa48618ace17d53539d"} Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.939240 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" event={"ID":"82b74254-de20-48a4-bfae-21afed7954cd","Type":"ContainerStarted","Data":"3efe394473f8f79631fddbedbe94c62d9c08dbf8c2151fc7b328233c928cd554"} Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.942774 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" event={"ID":"fd3cb745-7c4b-484b-93a5-7638dcd4d738","Type":"ContainerStarted","Data":"d676958352a72dcc02049875a6d2035305f9e16ae83a80e8127dc9c8f1d241b2"} Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.944042 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.150:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" podUID="82b74254-de20-48a4-bfae-21afed7954cd" Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.945895 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" event={"ID":"ea0ec30d-2c13-49fd-883b-767024ff632c","Type":"ContainerStarted","Data":"6c32e768d51115628c71f80f6b72d83b6fc75e37ad5e1755d22e164e7aae00fd"} Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.949046 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" event={"ID":"6848c014-0383-482f-9d46-232cc8fafbe1","Type":"ContainerStarted","Data":"11f3d78d45c9e40f1bf52321b584b08ea23b1976b5c9cd2bac67337b2fc4f45a"} Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.950825 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" event={"ID":"b54a05e7-9cfd-401d-8e5d-87c8ed5ad321","Type":"ContainerStarted","Data":"7113ae583797635f4fab83353ec8d14a347a29a964c5bc3279f26084c5be5ff2"} Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.952384 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" podUID="6848c014-0383-482f-9d46-232cc8fafbe1" Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.952654 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" podUID="b54a05e7-9cfd-401d-8e5d-87c8ed5ad321" Dec 03 00:44:42 crc kubenswrapper[4912]: I1203 00:44:42.968618 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" event={"ID":"b5544f39-6c3a-4737-a8e3-389a6e66ba40","Type":"ContainerStarted","Data":"0552def0934dc231253cc22b7d8add649e8b5759ad117ef5a77ce7a55640eadb"} Dec 03 00:44:42 crc kubenswrapper[4912]: E1203 00:44:42.969179 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" podUID="fd3cb745-7c4b-484b-93a5-7638dcd4d738" Dec 03 00:44:43 crc kubenswrapper[4912]: E1203 00:44:42.999524 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" podUID="b5544f39-6c3a-4737-a8e3-389a6e66ba40" Dec 03 00:44:43 crc kubenswrapper[4912]: I1203 00:44:42.999904 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" event={"ID":"8ddf2fc1-446c-4a26-987a-f67ec4c51455","Type":"ContainerStarted","Data":"2c281a6fbf77df336aa97916c31e5e71b2317d6d7f6bb525e22f0e400b820341"} Dec 03 00:44:43 crc kubenswrapper[4912]: I1203 00:44:43.010523 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" event={"ID":"de517f42-de01-4dde-ac22-1a651a38926b","Type":"ContainerStarted","Data":"c14a34f31f40e7fafb766514a52c7ea3a91408a43f3d3b3db94692b5b43a3344"} Dec 03 00:44:43 crc kubenswrapper[4912]: I1203 00:44:43.012168 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" event={"ID":"d2643e7f-8db2-4374-b0d9-05a2130795aa","Type":"ContainerStarted","Data":"42235e336f6afbe4e9ceb71739bf308aea8ae8ee9d4010a81a43dc7605e92ae4"} Dec 03 00:44:43 crc kubenswrapper[4912]: I1203 00:44:43.014693 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" event={"ID":"9b23d447-fdc8-47a1-af38-a569b6ad2a8f","Type":"ContainerStarted","Data":"e11934b75d217f19ff6d2db3774b3495d622c8550b52fe7cb5858be869af3c13"} Dec 03 00:44:43 crc kubenswrapper[4912]: I1203 00:44:43.016265 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" event={"ID":"2ab18e2a-3358-44e6-ac55-80b072619e73","Type":"ContainerStarted","Data":"e6e03d7d84fd67c3c296e0596bb1e4697080d16f5d648c1d696b69602ecd9180"} Dec 03 00:44:43 crc kubenswrapper[4912]: E1203 00:44:43.020236 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" podUID="2ab18e2a-3358-44e6-ac55-80b072619e73" Dec 03 00:44:43 crc kubenswrapper[4912]: I1203 00:44:43.028639 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" event={"ID":"5118703e-2fdb-41d6-a76e-ab5d3f6adb95","Type":"ContainerStarted","Data":"53c53534f3ec5d0e95d258e458bd3671b5e29d64501613488e1e08879060f26d"} Dec 03 00:44:43 crc kubenswrapper[4912]: I1203 00:44:43.699018 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:43 crc kubenswrapper[4912]: E1203 00:44:43.699322 4912 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:43 crc kubenswrapper[4912]: E1203 00:44:43.699382 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert podName:045ab366-9392-40f9-94f5-95c983f65176 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:47.69936322 +0000 UTC m=+1273.341383780 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert") pod "infra-operator-controller-manager-57548d458d-lzff2" (UID: "045ab366-9392-40f9-94f5-95c983f65176") : secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.060039 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" podUID="2ab18e2a-3358-44e6-ac55-80b072619e73" Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.060666 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.150:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" podUID="82b74254-de20-48a4-bfae-21afed7954cd" Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.060832 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" podUID="fd3cb745-7c4b-484b-93a5-7638dcd4d738" Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.063321 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" podUID="b5544f39-6c3a-4737-a8e3-389a6e66ba40" Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.063459 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" podUID="a7db7741-5afc-46bc-84c4-71c233ca9dde" Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.063577 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" podUID="b54a05e7-9cfd-401d-8e5d-87c8ed5ad321" Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.064298 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" podUID="6848c014-0383-482f-9d46-232cc8fafbe1" Dec 03 00:44:44 crc kubenswrapper[4912]: I1203 00:44:44.422147 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.422479 4912 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.422556 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert podName:3e85fc81-3fbd-4af2-8c98-89892a4f3579 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:48.422534733 +0000 UTC m=+1274.064555293 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" (UID: "3e85fc81-3fbd-4af2-8c98-89892a4f3579") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:44 crc kubenswrapper[4912]: I1203 00:44:44.952386 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.952630 4912 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.953068 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:48.953044066 +0000 UTC m=+1274.595064626 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "metrics-server-cert" not found Dec 03 00:44:44 crc kubenswrapper[4912]: I1203 00:44:44.952977 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.953104 4912 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 00:44:44 crc kubenswrapper[4912]: E1203 00:44:44.953259 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:48.953235201 +0000 UTC m=+1274.595255761 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "webhook-server-cert" not found Dec 03 00:44:47 crc kubenswrapper[4912]: I1203 00:44:47.708950 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:47 crc kubenswrapper[4912]: E1203 00:44:47.709359 4912 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:47 crc kubenswrapper[4912]: E1203 00:44:47.709498 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert podName:045ab366-9392-40f9-94f5-95c983f65176 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:55.70942338 +0000 UTC m=+1281.351443950 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert") pod "infra-operator-controller-manager-57548d458d-lzff2" (UID: "045ab366-9392-40f9-94f5-95c983f65176") : secret "infra-operator-webhook-server-cert" not found Dec 03 00:44:48 crc kubenswrapper[4912]: I1203 00:44:48.427927 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:48 crc kubenswrapper[4912]: E1203 00:44:48.428440 4912 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:48 crc kubenswrapper[4912]: E1203 00:44:48.428557 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert podName:3e85fc81-3fbd-4af2-8c98-89892a4f3579 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:56.428531563 +0000 UTC m=+1282.070552123 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" (UID: "3e85fc81-3fbd-4af2-8c98-89892a4f3579") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 00:44:49 crc kubenswrapper[4912]: I1203 00:44:49.043027 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:49 crc kubenswrapper[4912]: I1203 00:44:49.043603 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:49 crc kubenswrapper[4912]: E1203 00:44:49.043321 4912 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 00:44:49 crc kubenswrapper[4912]: E1203 00:44:49.043749 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:57.043712405 +0000 UTC m=+1282.685732965 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "metrics-server-cert" not found Dec 03 00:44:49 crc kubenswrapper[4912]: E1203 00:44:49.043787 4912 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 00:44:49 crc kubenswrapper[4912]: E1203 00:44:49.043902 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:44:57.043868419 +0000 UTC m=+1282.685889210 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "webhook-server-cert" not found Dec 03 00:44:54 crc kubenswrapper[4912]: E1203 00:44:54.293670 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 03 00:44:54 crc kubenswrapper[4912]: E1203 00:44:54.295320 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dm8cc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-jskq6_openstack-operators(1df68f98-9a63-43ca-8e47-5c34c1e52581): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:44:54 crc kubenswrapper[4912]: E1203 00:44:54.941300 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 03 00:44:54 crc kubenswrapper[4912]: E1203 00:44:54.942316 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rhc5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-rb4z7_openstack-operators(27605b80-0a7f-4b6b-9729-d0b6eaa0a74b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:44:55 crc kubenswrapper[4912]: E1203 00:44:55.516063 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 03 00:44:55 crc kubenswrapper[4912]: E1203 00:44:55.516243 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s57k2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-tqphq_openstack-operators(5118703e-2fdb-41d6-a76e-ab5d3f6adb95): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:44:55 crc kubenswrapper[4912]: I1203 00:44:55.785400 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:55 crc kubenswrapper[4912]: I1203 00:44:55.797264 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/045ab366-9392-40f9-94f5-95c983f65176-cert\") pod \"infra-operator-controller-manager-57548d458d-lzff2\" (UID: \"045ab366-9392-40f9-94f5-95c983f65176\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:55 crc kubenswrapper[4912]: I1203 00:44:55.862372 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:44:56 crc kubenswrapper[4912]: I1203 00:44:56.506285 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:56 crc kubenswrapper[4912]: I1203 00:44:56.512104 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e85fc81-3fbd-4af2-8c98-89892a4f3579-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb\" (UID: \"3e85fc81-3fbd-4af2-8c98-89892a4f3579\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:56 crc kubenswrapper[4912]: I1203 00:44:56.746998 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:44:57 crc kubenswrapper[4912]: I1203 00:44:57.117566 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:57 crc kubenswrapper[4912]: I1203 00:44:57.117638 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:44:57 crc kubenswrapper[4912]: E1203 00:44:57.117856 4912 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 00:44:57 crc kubenswrapper[4912]: E1203 00:44:57.117949 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:45:13.117929752 +0000 UTC m=+1298.759950312 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "metrics-server-cert" not found Dec 03 00:44:57 crc kubenswrapper[4912]: E1203 00:44:57.117865 4912 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 00:44:57 crc kubenswrapper[4912]: E1203 00:44:57.118376 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs podName:10227b7f-ff98-4f4a-862f-3ef78b4e7737 nodeName:}" failed. No retries permitted until 2025-12-03 00:45:13.118323723 +0000 UTC m=+1298.760344283 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs") pod "openstack-operator-controller-manager-86464467f8-sq576" (UID: "10227b7f-ff98-4f4a-862f-3ef78b4e7737") : secret "webhook-server-cert" not found Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.169166 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b"] Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.171343 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.173751 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.174086 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.183241 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b"] Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.290018 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5be2251-7d9d-4360-8e88-d71bc8458cf8-config-volume\") pod \"collect-profiles-29412045-hh55b\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.290156 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gskf\" (UniqueName: \"kubernetes.io/projected/a5be2251-7d9d-4360-8e88-d71bc8458cf8-kube-api-access-6gskf\") pod \"collect-profiles-29412045-hh55b\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.290230 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5be2251-7d9d-4360-8e88-d71bc8458cf8-secret-volume\") pod \"collect-profiles-29412045-hh55b\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.391719 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5be2251-7d9d-4360-8e88-d71bc8458cf8-config-volume\") pod \"collect-profiles-29412045-hh55b\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.391854 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gskf\" (UniqueName: \"kubernetes.io/projected/a5be2251-7d9d-4360-8e88-d71bc8458cf8-kube-api-access-6gskf\") pod \"collect-profiles-29412045-hh55b\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.391926 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5be2251-7d9d-4360-8e88-d71bc8458cf8-secret-volume\") pod \"collect-profiles-29412045-hh55b\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.394060 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5be2251-7d9d-4360-8e88-d71bc8458cf8-config-volume\") pod \"collect-profiles-29412045-hh55b\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.414250 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5be2251-7d9d-4360-8e88-d71bc8458cf8-secret-volume\") pod \"collect-profiles-29412045-hh55b\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.417919 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gskf\" (UniqueName: \"kubernetes.io/projected/a5be2251-7d9d-4360-8e88-d71bc8458cf8-kube-api-access-6gskf\") pod \"collect-profiles-29412045-hh55b\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:00 crc kubenswrapper[4912]: I1203 00:45:00.496402 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.460969 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.462231 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2c65n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-bdgdj_openstack-operators(de517f42-de01-4dde-ac22-1a651a38926b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.526057 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.526671 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dm8cc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-jskq6_openstack-operators(1df68f98-9a63-43ca-8e47-5c34c1e52581): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.528274 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.528344 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" podUID="1df68f98-9a63-43ca-8e47-5c34c1e52581" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.528560 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rhc5w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-rb4z7_openstack-operators(27605b80-0a7f-4b6b-9729-d0b6eaa0a74b): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.529921 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" podUID="27605b80-0a7f-4b6b-9729-d0b6eaa0a74b" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.919039 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:d7a3730895772e1892ec7c4128521360742116d666d304c68e4122f6a24225b7: Get \"http://38.102.83.150:5001/v2/openstack-k8s-operators/telemetry-operator/blobs/sha256:d7a3730895772e1892ec7c4128521360742116d666d304c68e4122f6a24225b7\": context canceled" image="38.102.83.150:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.919619 4912 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = reading blob sha256:d7a3730895772e1892ec7c4128521360742116d666d304c68e4122f6a24225b7: Get \"http://38.102.83.150:5001/v2/openstack-k8s-operators/telemetry-operator/blobs/sha256:d7a3730895772e1892ec7c4128521360742116d666d304c68e4122f6a24225b7\": context canceled" image="38.102.83.150:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9" Dec 03 00:45:06 crc kubenswrapper[4912]: E1203 00:45:06.919835 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.150:5001/openstack-k8s-operators/telemetry-operator:c3240f18e0dac2c05c89e1470dd21f967fd0cbe9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kfnmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5d66d5c957-xgv9t_openstack-operators(82b74254-de20-48a4-bfae-21afed7954cd): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:d7a3730895772e1892ec7c4128521360742116d666d304c68e4122f6a24225b7: Get \"http://38.102.83.150:5001/v2/openstack-k8s-operators/telemetry-operator/blobs/sha256:d7a3730895772e1892ec7c4128521360742116d666d304c68e4122f6a24225b7\": context canceled" logger="UnhandledError" Dec 03 00:45:07 crc kubenswrapper[4912]: I1203 00:45:07.027413 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-lzff2"] Dec 03 00:45:07 crc kubenswrapper[4912]: I1203 00:45:07.149503 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b"] Dec 03 00:45:07 crc kubenswrapper[4912]: I1203 00:45:07.166621 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb"] Dec 03 00:45:07 crc kubenswrapper[4912]: W1203 00:45:07.239727 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5be2251_7d9d_4360_8e88_d71bc8458cf8.slice/crio-eb227f65d483a8df53a0cec0c563ec10f325e824a0b007accf1a03a10d08433e WatchSource:0}: Error finding container eb227f65d483a8df53a0cec0c563ec10f325e824a0b007accf1a03a10d08433e: Status 404 returned error can't find the container with id eb227f65d483a8df53a0cec0c563ec10f325e824a0b007accf1a03a10d08433e Dec 03 00:45:07 crc kubenswrapper[4912]: W1203 00:45:07.241756 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e85fc81_3fbd_4af2_8c98_89892a4f3579.slice/crio-aed78aaa7348833e0ee61514e113f3728889bd7fe586fca1ea4780fa185de645 WatchSource:0}: Error finding container aed78aaa7348833e0ee61514e113f3728889bd7fe586fca1ea4780fa185de645: Status 404 returned error can't find the container with id aed78aaa7348833e0ee61514e113f3728889bd7fe586fca1ea4780fa185de645 Dec 03 00:45:07 crc kubenswrapper[4912]: I1203 00:45:07.338064 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" event={"ID":"8ddf2fc1-446c-4a26-987a-f67ec4c51455","Type":"ContainerStarted","Data":"27938bf13a4413a50b5f6ac1676992cbe2e29310ea2417f68f1c8f6c64547a4f"} Dec 03 00:45:07 crc kubenswrapper[4912]: I1203 00:45:07.341885 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" event={"ID":"baaa7298-1635-4de5-af87-68db95102082","Type":"ContainerStarted","Data":"a0d40da54293267816ea37dd3d56db06ab704adbc7baa95cd44c715d739ba53b"} Dec 03 00:45:07 crc kubenswrapper[4912]: I1203 00:45:07.350908 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" event={"ID":"56cdea32-a9a4-410b-9a9c-c237746b0582","Type":"ContainerStarted","Data":"81f0c4848155b420273c5ed6c5bc4fc903bedecb16b220535a27543534e3c565"} Dec 03 00:45:07 crc kubenswrapper[4912]: I1203 00:45:07.358691 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" event={"ID":"ea0ec30d-2c13-49fd-883b-767024ff632c","Type":"ContainerStarted","Data":"e01d15ec9583fe79c544b24b5ad42394241e8343b0fbfec22833c17dbb5cafe9"} Dec 03 00:45:07 crc kubenswrapper[4912]: I1203 00:45:07.361546 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" event={"ID":"a5be2251-7d9d-4360-8e88-d71bc8458cf8","Type":"ContainerStarted","Data":"eb227f65d483a8df53a0cec0c563ec10f325e824a0b007accf1a03a10d08433e"} Dec 03 00:45:07 crc kubenswrapper[4912]: I1203 00:45:07.363885 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" event={"ID":"3e85fc81-3fbd-4af2-8c98-89892a4f3579","Type":"ContainerStarted","Data":"aed78aaa7348833e0ee61514e113f3728889bd7fe586fca1ea4780fa185de645"} Dec 03 00:45:07 crc kubenswrapper[4912]: W1203 00:45:07.934338 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod045ab366_9392_40f9_94f5_95c983f65176.slice/crio-caf24772a7a3f2121e265fa4f7985ab0ae06eabccd07fc4ad04cb7094db6e508 WatchSource:0}: Error finding container caf24772a7a3f2121e265fa4f7985ab0ae06eabccd07fc4ad04cb7094db6e508: Status 404 returned error can't find the container with id caf24772a7a3f2121e265fa4f7985ab0ae06eabccd07fc4ad04cb7094db6e508 Dec 03 00:45:08 crc kubenswrapper[4912]: I1203 00:45:08.407346 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" event={"ID":"7b515331-ceab-4f45-9880-719c72dfcc4c","Type":"ContainerStarted","Data":"f5732f9b2b4f1b2e34b0c629ebe8389af239ec1308a3fc7dcb0164d7040f2ff8"} Dec 03 00:45:08 crc kubenswrapper[4912]: I1203 00:45:08.436122 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" event={"ID":"045ab366-9392-40f9-94f5-95c983f65176","Type":"ContainerStarted","Data":"caf24772a7a3f2121e265fa4f7985ab0ae06eabccd07fc4ad04cb7094db6e508"} Dec 03 00:45:08 crc kubenswrapper[4912]: I1203 00:45:08.443389 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" event={"ID":"7ec29496-d0ac-42a2-acea-7037dfde41fb","Type":"ContainerStarted","Data":"19c0fc53a67c60f8956f0a86c95642c2cff50b943c1d9520b6c460b2061f8c95"} Dec 03 00:45:09 crc kubenswrapper[4912]: I1203 00:45:09.460390 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" event={"ID":"d2643e7f-8db2-4374-b0d9-05a2130795aa","Type":"ContainerStarted","Data":"c87fdc707066a50c57c5dd950c79224104a768debcc60c70424490ae28b34aed"} Dec 03 00:45:09 crc kubenswrapper[4912]: I1203 00:45:09.466610 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" event={"ID":"9b23d447-fdc8-47a1-af38-a569b6ad2a8f","Type":"ContainerStarted","Data":"79324d8335b8714e581735fcf1f940969d78e6a7e508ea3276e521ad4975f524"} Dec 03 00:45:09 crc kubenswrapper[4912]: I1203 00:45:09.471659 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" event={"ID":"70a932ec-65ac-4616-870b-9297bd5d0a28","Type":"ContainerStarted","Data":"6730c9d2a5f763fec0139184ac85e6b7728bb227e7e64a0e20360941e24c3184"} Dec 03 00:45:12 crc kubenswrapper[4912]: I1203 00:45:12.501125 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" event={"ID":"a5be2251-7d9d-4360-8e88-d71bc8458cf8","Type":"ContainerStarted","Data":"5bdc7bf2168fd20ce5de41930c506e045710fcf1a64f7e19bee6006c69e5b0a3"} Dec 03 00:45:12 crc kubenswrapper[4912]: I1203 00:45:12.527505 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" podStartSLOduration=12.527481588 podStartE2EDuration="12.527481588s" podCreationTimestamp="2025-12-03 00:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:45:12.521762797 +0000 UTC m=+1298.163783367" watchObservedRunningTime="2025-12-03 00:45:12.527481588 +0000 UTC m=+1298.169502148" Dec 03 00:45:13 crc kubenswrapper[4912]: I1203 00:45:13.220049 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:45:13 crc kubenswrapper[4912]: I1203 00:45:13.220694 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:45:13 crc kubenswrapper[4912]: I1203 00:45:13.229810 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-metrics-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:45:13 crc kubenswrapper[4912]: I1203 00:45:13.246388 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/10227b7f-ff98-4f4a-862f-3ef78b4e7737-webhook-certs\") pod \"openstack-operator-controller-manager-86464467f8-sq576\" (UID: \"10227b7f-ff98-4f4a-862f-3ef78b4e7737\") " pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:45:13 crc kubenswrapper[4912]: I1203 00:45:13.442025 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:45:13 crc kubenswrapper[4912]: I1203 00:45:13.511575 4912 generic.go:334] "Generic (PLEG): container finished" podID="a5be2251-7d9d-4360-8e88-d71bc8458cf8" containerID="5bdc7bf2168fd20ce5de41930c506e045710fcf1a64f7e19bee6006c69e5b0a3" exitCode=0 Dec 03 00:45:13 crc kubenswrapper[4912]: I1203 00:45:13.511678 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" event={"ID":"a5be2251-7d9d-4360-8e88-d71bc8458cf8","Type":"ContainerDied","Data":"5bdc7bf2168fd20ce5de41930c506e045710fcf1a64f7e19bee6006c69e5b0a3"} Dec 03 00:45:18 crc kubenswrapper[4912]: I1203 00:45:18.078036 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:45:18 crc kubenswrapper[4912]: I1203 00:45:18.079099 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.140899 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.142545 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gtt7v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-lzff2_openstack-operators(045ab366-9392-40f9-94f5-95c983f65176): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.751759 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.752366 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9hfgk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb_openstack-operators(3e85fc81-3fbd-4af2-8c98-89892a4f3579): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.776995 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.777267 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hc7qv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-fxhh2_openstack-operators(8ddf2fc1-446c-4a26-987a-f67ec4c51455): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.777671 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.777772 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kfnmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5d66d5c957-xgv9t_openstack-operators(82b74254-de20-48a4-bfae-21afed7954cd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.778785 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" podUID="8ddf2fc1-446c-4a26-987a-f67ec4c51455" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.780171 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:d7a3730895772e1892ec7c4128521360742116d666d304c68e4122f6a24225b7: Get \\\"http://38.102.83.150:5001/v2/openstack-k8s-operators/telemetry-operator/blobs/sha256:d7a3730895772e1892ec7c4128521360742116d666d304c68e4122f6a24225b7\\\": context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" podUID="82b74254-de20-48a4-bfae-21afed7954cd" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.831932 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.832638 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nqtpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-jqzw7_openstack-operators(ea0ec30d-2c13-49fd-883b-767024ff632c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.833973 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" podUID="ea0ec30d-2c13-49fd-883b-767024ff632c" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.838055 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.838162 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ltcqf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-dtgrm_openstack-operators(baaa7298-1635-4de5-af87-68db95102082): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:26 crc kubenswrapper[4912]: E1203 00:45:26.840250 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" podUID="baaa7298-1635-4de5-af87-68db95102082" Dec 03 00:45:26 crc kubenswrapper[4912]: I1203 00:45:26.931358 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.131735 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5be2251-7d9d-4360-8e88-d71bc8458cf8-secret-volume\") pod \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.131926 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5be2251-7d9d-4360-8e88-d71bc8458cf8-config-volume\") pod \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.132025 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gskf\" (UniqueName: \"kubernetes.io/projected/a5be2251-7d9d-4360-8e88-d71bc8458cf8-kube-api-access-6gskf\") pod \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\" (UID: \"a5be2251-7d9d-4360-8e88-d71bc8458cf8\") " Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.133112 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5be2251-7d9d-4360-8e88-d71bc8458cf8-config-volume" (OuterVolumeSpecName: "config-volume") pod "a5be2251-7d9d-4360-8e88-d71bc8458cf8" (UID: "a5be2251-7d9d-4360-8e88-d71bc8458cf8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.140735 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5be2251-7d9d-4360-8e88-d71bc8458cf8-kube-api-access-6gskf" (OuterVolumeSpecName: "kube-api-access-6gskf") pod "a5be2251-7d9d-4360-8e88-d71bc8458cf8" (UID: "a5be2251-7d9d-4360-8e88-d71bc8458cf8"). InnerVolumeSpecName "kube-api-access-6gskf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.149564 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5be2251-7d9d-4360-8e88-d71bc8458cf8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a5be2251-7d9d-4360-8e88-d71bc8458cf8" (UID: "a5be2251-7d9d-4360-8e88-d71bc8458cf8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.234586 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5be2251-7d9d-4360-8e88-d71bc8458cf8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.234628 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gskf\" (UniqueName: \"kubernetes.io/projected/a5be2251-7d9d-4360-8e88-d71bc8458cf8-kube-api-access-6gskf\") on node \"crc\" DevicePath \"\"" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.234642 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5be2251-7d9d-4360-8e88-d71bc8458cf8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.411354 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.411788 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7jg95,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-626ss_openstack-operators(7b515331-ceab-4f45-9880-719c72dfcc4c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.411879 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.412006 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9svhq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-5n67v_openstack-operators(7ec29496-d0ac-42a2-acea-7037dfde41fb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.413221 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" podUID="7ec29496-d0ac-42a2-acea-7037dfde41fb" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.413292 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" podUID="7b515331-ceab-4f45-9880-719c72dfcc4c" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.440269 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.440516 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f8ccx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-llhvt_openstack-operators(b54a05e7-9cfd-401d-8e5d-87c8ed5ad321): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.441691 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" podUID="b54a05e7-9cfd-401d-8e5d-87c8ed5ad321" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.449272 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.449505 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vczzz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-twd26_openstack-operators(56cdea32-a9a4-410b-9a9c-c237746b0582): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.450372 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.450649 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8l7mh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-txtqb_openstack-operators(9b23d447-fdc8-47a1-af38-a569b6ad2a8f): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.450800 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.450952 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8g6fg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-g4n8g_openstack-operators(d2643e7f-8db2-4374-b0d9-05a2130795aa): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.451142 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" podUID="56cdea32-a9a4-410b-9a9c-c237746b0582" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.451976 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" podUID="9b23d447-fdc8-47a1-af38-a569b6ad2a8f" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.452122 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" podUID="d2643e7f-8db2-4374-b0d9-05a2130795aa" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.452814 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.453191 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rpltn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-ds5mt_openstack-operators(70a932ec-65ac-4616-870b-9297bd5d0a28): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 03 00:45:27 crc kubenswrapper[4912]: E1203 00:45:27.454350 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" podUID="70a932ec-65ac-4616-870b-9297bd5d0a28" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.686280 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" event={"ID":"1df68f98-9a63-43ca-8e47-5c34c1e52581","Type":"ContainerStarted","Data":"f4e4128858c1277994ae162d7111d60dd12bb58a136a1c3b902047de1550382c"} Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.702702 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" event={"ID":"a5be2251-7d9d-4360-8e88-d71bc8458cf8","Type":"ContainerDied","Data":"eb227f65d483a8df53a0cec0c563ec10f325e824a0b007accf1a03a10d08433e"} Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.702781 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb227f65d483a8df53a0cec0c563ec10f325e824a0b007accf1a03a10d08433e" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.702781 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.704154 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.704795 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.704830 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.710562 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.711109 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" Dec 03 00:45:27 crc kubenswrapper[4912]: I1203 00:45:27.717621 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.146443 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86464467f8-sq576"] Dec 03 00:45:28 crc kubenswrapper[4912]: W1203 00:45:28.254915 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10227b7f_ff98_4f4a_862f_3ef78b4e7737.slice/crio-533f5d8875e66e5fa0929045e1a18cc654ed05073a088d26b36d4bf71b4caa24 WatchSource:0}: Error finding container 533f5d8875e66e5fa0929045e1a18cc654ed05073a088d26b36d4bf71b4caa24: Status 404 returned error can't find the container with id 533f5d8875e66e5fa0929045e1a18cc654ed05073a088d26b36d4bf71b4caa24 Dec 03 00:45:28 crc kubenswrapper[4912]: E1203 00:45:28.329572 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" podUID="5118703e-2fdb-41d6-a76e-ab5d3f6adb95" Dec 03 00:45:28 crc kubenswrapper[4912]: E1203 00:45:28.361208 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" podUID="de517f42-de01-4dde-ac22-1a651a38926b" Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.728881 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" event={"ID":"1df68f98-9a63-43ca-8e47-5c34c1e52581","Type":"ContainerStarted","Data":"a0bf0ce7b03e13c531a1edbf9c99ef39291542747c4110da6875382bb7142c31"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.729215 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.735625 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" event={"ID":"6848c014-0383-482f-9d46-232cc8fafbe1","Type":"ContainerStarted","Data":"11197583d96a53f9dadc9850fb6f8bc8c0845a952569e4a864469a2a169eda6e"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.735656 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" event={"ID":"6848c014-0383-482f-9d46-232cc8fafbe1","Type":"ContainerStarted","Data":"b6aa3f8225ebdd53cd53d6ae6747309cb14090b0dbafb7b16739ccd97ca02286"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.740370 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" event={"ID":"a7db7741-5afc-46bc-84c4-71c233ca9dde","Type":"ContainerStarted","Data":"9b63bee50d09ce18d9cca967335aac327da05f8a762a5e76becd103d0d18c819"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.754921 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" event={"ID":"2ab18e2a-3358-44e6-ac55-80b072619e73","Type":"ContainerStarted","Data":"71c4b67bf3aa8811b3ef3f6b22589e9ca5b734d10504eda4425f6272e5a0aca4"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.755028 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" podStartSLOduration=23.74272733 podStartE2EDuration="49.75500918s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.534009681 +0000 UTC m=+1268.176030231" lastFinishedPulling="2025-12-03 00:45:08.546291521 +0000 UTC m=+1294.188312081" observedRunningTime="2025-12-03 00:45:28.751070166 +0000 UTC m=+1314.393090726" watchObservedRunningTime="2025-12-03 00:45:28.75500918 +0000 UTC m=+1314.397029740" Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.762787 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" event={"ID":"5118703e-2fdb-41d6-a76e-ab5d3f6adb95","Type":"ContainerStarted","Data":"876ec31829d64771df9625ef9d8ce6434dc4f688e169b7a07ff9d1550e42205d"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.775407 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" event={"ID":"b5544f39-6c3a-4737-a8e3-389a6e66ba40","Type":"ContainerStarted","Data":"8c5d3ed9e8819cac8775a52b68e3cb6d88e2cd5c5fa3880414baa9fc6cc2a949"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.780714 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" event={"ID":"fd3cb745-7c4b-484b-93a5-7638dcd4d738","Type":"ContainerStarted","Data":"4e4069cfa3fd656b5744eb3bb04b16cd12fae130d7ef89d1e97a796bc6a61d7a"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.786909 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" event={"ID":"10227b7f-ff98-4f4a-862f-3ef78b4e7737","Type":"ContainerStarted","Data":"533f5d8875e66e5fa0929045e1a18cc654ed05073a088d26b36d4bf71b4caa24"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.819795 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" event={"ID":"d2643e7f-8db2-4374-b0d9-05a2130795aa","Type":"ContainerStarted","Data":"563c6b7c2901992e3ab7d67c729574f0673af4243fddfbf13155180e49045fa3"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.841064 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" event={"ID":"de517f42-de01-4dde-ac22-1a651a38926b","Type":"ContainerStarted","Data":"e25879f4369b2b472bda9d3df36b6389b532a7034fc85b0d635df455fbfd3b89"} Dec 03 00:45:28 crc kubenswrapper[4912]: I1203 00:45:28.856627 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" event={"ID":"27605b80-0a7f-4b6b-9729-d0b6eaa0a74b","Type":"ContainerStarted","Data":"75a88e50ad620588646721a73e05c8d64a496db27fefff3d3186e7b137388f00"} Dec 03 00:45:28 crc kubenswrapper[4912]: E1203 00:45:28.969338 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" podUID="3e85fc81-3fbd-4af2-8c98-89892a4f3579" Dec 03 00:45:29 crc kubenswrapper[4912]: E1203 00:45:29.737392 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" podUID="045ab366-9392-40f9-94f5-95c983f65176" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.880705 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" event={"ID":"27605b80-0a7f-4b6b-9729-d0b6eaa0a74b","Type":"ContainerStarted","Data":"35b47288eb6c498a84ea10f3ce271a1f919a834d977ae9674e22fa0b71b5f76b"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.880830 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.892945 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" event={"ID":"10227b7f-ff98-4f4a-862f-3ef78b4e7737","Type":"ContainerStarted","Data":"e0f5eeb0ba9049ba871bda60ce8bc657fc50ec0c07067de85937662a73a67139"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.893202 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.908970 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" event={"ID":"fd3cb745-7c4b-484b-93a5-7638dcd4d738","Type":"ContainerStarted","Data":"ab983ba6090f95347d416235f48172f85cbacddd29952584619e9067588efa5d"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.909667 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.913946 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" event={"ID":"7ec29496-d0ac-42a2-acea-7037dfde41fb","Type":"ContainerStarted","Data":"281999fd113e8bd7d72c07222e4d2ab4d9c37bba62e662156a9806cc967eeac2"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.920771 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" podStartSLOduration=22.792032499 podStartE2EDuration="50.920756158s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.491535518 +0000 UTC m=+1268.133556088" lastFinishedPulling="2025-12-03 00:45:10.620259187 +0000 UTC m=+1296.262279747" observedRunningTime="2025-12-03 00:45:29.91663771 +0000 UTC m=+1315.558658280" watchObservedRunningTime="2025-12-03 00:45:29.920756158 +0000 UTC m=+1315.562776718" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.922690 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" event={"ID":"9b23d447-fdc8-47a1-af38-a569b6ad2a8f","Type":"ContainerStarted","Data":"a9659a50097c58050c1a626b0f51373eae9764f0763fbb01910b4f9aac153782"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.926363 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.930258 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" event={"ID":"a7db7741-5afc-46bc-84c4-71c233ca9dde","Type":"ContainerStarted","Data":"6180a1e02d8b44e3662c95f4a93324a404818b9da3e43b4e4861e6a63f53167c"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.931483 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.931825 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.939952 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" event={"ID":"2ab18e2a-3358-44e6-ac55-80b072619e73","Type":"ContainerStarted","Data":"b6ceaff4b730be5d57714b828ce00f41b7c510985bca2d6564d55260e7be0f80"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.941100 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.943900 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" podStartSLOduration=5.037945542 podStartE2EDuration="49.943885176s" podCreationTimestamp="2025-12-03 00:44:40 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.534396592 +0000 UTC m=+1268.176417162" lastFinishedPulling="2025-12-03 00:45:27.440336236 +0000 UTC m=+1313.082356796" observedRunningTime="2025-12-03 00:45:29.939101401 +0000 UTC m=+1315.581121971" watchObservedRunningTime="2025-12-03 00:45:29.943885176 +0000 UTC m=+1315.585905736" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.952782 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" event={"ID":"7b515331-ceab-4f45-9880-719c72dfcc4c","Type":"ContainerStarted","Data":"8a0d742bab700a811ab457ea457c32f3d7d86926f6c02f38a8ff4a79ee87a750"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.955527 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.957674 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.963936 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" event={"ID":"70a932ec-65ac-4616-870b-9297bd5d0a28","Type":"ContainerStarted","Data":"b16a503299b89b430a51c04f9a55526f26dc0b6b2b4b08cc2993fdfa836b3eb8"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.965095 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.969224 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" event={"ID":"b5544f39-6c3a-4737-a8e3-389a6e66ba40","Type":"ContainerStarted","Data":"df01fe570cec7247b19bd54f54913a1d093de31d71dd0e6fdc896cee99697317"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.969400 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.970034 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.970856 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" event={"ID":"3e85fc81-3fbd-4af2-8c98-89892a4f3579","Type":"ContainerStarted","Data":"02a3c17648c584702935e238168806c07dbd83df31386893096f6fa1e21620bc"} Dec 03 00:45:29 crc kubenswrapper[4912]: E1203 00:45:29.972070 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" podUID="3e85fc81-3fbd-4af2-8c98-89892a4f3579" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.974980 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" event={"ID":"56cdea32-a9a4-410b-9a9c-c237746b0582","Type":"ContainerStarted","Data":"bf7bbdf04e1584bb34c9d4e73838959dd9bd978cf541e05de15478fe1eecefb8"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.975384 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.977992 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" event={"ID":"ea0ec30d-2c13-49fd-883b-767024ff632c","Type":"ContainerStarted","Data":"17593730f6f4fdbdcf6e0b744ceac9db658e408c2248fe2a0b7b840a74e5a061"} Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.978340 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.985811 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.985954 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.991186 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-5n67v" podStartSLOduration=25.89348705 podStartE2EDuration="49.991157439s" podCreationTimestamp="2025-12-03 00:44:40 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.409212473 +0000 UTC m=+1268.051233033" lastFinishedPulling="2025-12-03 00:45:06.506882862 +0000 UTC m=+1292.148903422" observedRunningTime="2025-12-03 00:45:29.985660885 +0000 UTC m=+1315.627681445" watchObservedRunningTime="2025-12-03 00:45:29.991157439 +0000 UTC m=+1315.633177999" Dec 03 00:45:29 crc kubenswrapper[4912]: I1203 00:45:29.994799 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" event={"ID":"8ddf2fc1-446c-4a26-987a-f67ec4c51455","Type":"ContainerStarted","Data":"17ee595fee711c475129423ba47e7cbf3d9df2e6b5dc35457fc99c1391e59157"} Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.034378 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" event={"ID":"045ab366-9392-40f9-94f5-95c983f65176","Type":"ContainerStarted","Data":"b4b0f693841acd4ca4aeb72787fc2e73463d5e2783112b7fe19e04047132463b"} Dec 03 00:45:30 crc kubenswrapper[4912]: E1203 00:45:30.051048 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" podUID="045ab366-9392-40f9-94f5-95c983f65176" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.056569 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" podStartSLOduration=50.056542019 podStartE2EDuration="50.056542019s" podCreationTimestamp="2025-12-03 00:44:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:45:30.024478416 +0000 UTC m=+1315.666498976" watchObservedRunningTime="2025-12-03 00:45:30.056542019 +0000 UTC m=+1315.698562579" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.082799 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-txtqb" podStartSLOduration=27.090265115 podStartE2EDuration="51.082767408s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.534284939 +0000 UTC m=+1268.176305499" lastFinishedPulling="2025-12-03 00:45:06.526787232 +0000 UTC m=+1292.168807792" observedRunningTime="2025-12-03 00:45:30.064031826 +0000 UTC m=+1315.706052386" watchObservedRunningTime="2025-12-03 00:45:30.082767408 +0000 UTC m=+1315.724787968" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.096840 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" event={"ID":"baaa7298-1635-4de5-af87-68db95102082","Type":"ContainerStarted","Data":"0299119227ca1e493387c1681c41e498b4bf99b3a7b816fbce28cbd6466fd237"} Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.098005 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.098058 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.109171 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.150513 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-626ss" podStartSLOduration=27.177532337 podStartE2EDuration="51.1504876s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.533998611 +0000 UTC m=+1268.176019171" lastFinishedPulling="2025-12-03 00:45:06.506953874 +0000 UTC m=+1292.148974434" observedRunningTime="2025-12-03 00:45:30.135990198 +0000 UTC m=+1315.778010748" watchObservedRunningTime="2025-12-03 00:45:30.1504876 +0000 UTC m=+1315.792508170" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.164224 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" podStartSLOduration=5.276535062 podStartE2EDuration="50.16419991s" podCreationTimestamp="2025-12-03 00:44:40 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.552582356 +0000 UTC m=+1268.194602916" lastFinishedPulling="2025-12-03 00:45:27.440247214 +0000 UTC m=+1313.082267764" observedRunningTime="2025-12-03 00:45:30.098570835 +0000 UTC m=+1315.740591405" watchObservedRunningTime="2025-12-03 00:45:30.16419991 +0000 UTC m=+1315.806220470" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.224346 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" podStartSLOduration=6.374937136 podStartE2EDuration="51.224317461s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.559486544 +0000 UTC m=+1268.201507104" lastFinishedPulling="2025-12-03 00:45:27.408866829 +0000 UTC m=+1313.050887429" observedRunningTime="2025-12-03 00:45:30.219844563 +0000 UTC m=+1315.861865133" watchObservedRunningTime="2025-12-03 00:45:30.224317461 +0000 UTC m=+1315.866338021" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.253930 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-jqzw7" podStartSLOduration=26.442171294 podStartE2EDuration="50.25391104s" podCreationTimestamp="2025-12-03 00:44:40 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.663492027 +0000 UTC m=+1268.305512587" lastFinishedPulling="2025-12-03 00:45:06.475231783 +0000 UTC m=+1292.117252333" observedRunningTime="2025-12-03 00:45:30.250719565 +0000 UTC m=+1315.892740135" watchObservedRunningTime="2025-12-03 00:45:30.25391104 +0000 UTC m=+1315.895931600" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.279093 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-ds5mt" podStartSLOduration=26.522723221 podStartE2EDuration="51.279072942s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:41.750540801 +0000 UTC m=+1267.392561361" lastFinishedPulling="2025-12-03 00:45:06.506890522 +0000 UTC m=+1292.148911082" observedRunningTime="2025-12-03 00:45:30.274370907 +0000 UTC m=+1315.916391467" watchObservedRunningTime="2025-12-03 00:45:30.279072942 +0000 UTC m=+1315.921093502" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.309810 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-twd26" podStartSLOduration=26.349191451 podStartE2EDuration="51.309795529s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:41.514625015 +0000 UTC m=+1267.156645565" lastFinishedPulling="2025-12-03 00:45:06.475229083 +0000 UTC m=+1292.117249643" observedRunningTime="2025-12-03 00:45:30.304444139 +0000 UTC m=+1315.946464699" watchObservedRunningTime="2025-12-03 00:45:30.309795529 +0000 UTC m=+1315.951816089" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.363413 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-fxhh2" podStartSLOduration=27.358741606 podStartE2EDuration="51.363387769s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.502193578 +0000 UTC m=+1268.144214138" lastFinishedPulling="2025-12-03 00:45:06.506839731 +0000 UTC m=+1292.148860301" observedRunningTime="2025-12-03 00:45:30.350314684 +0000 UTC m=+1315.992335264" watchObservedRunningTime="2025-12-03 00:45:30.363387769 +0000 UTC m=+1316.005408329" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.385981 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" podStartSLOduration=5.659807838 podStartE2EDuration="50.385959742s" podCreationTimestamp="2025-12-03 00:44:40 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.723178578 +0000 UTC m=+1268.365199148" lastFinishedPulling="2025-12-03 00:45:27.449330482 +0000 UTC m=+1313.091351052" observedRunningTime="2025-12-03 00:45:30.379945704 +0000 UTC m=+1316.021966264" watchObservedRunningTime="2025-12-03 00:45:30.385959742 +0000 UTC m=+1316.027980302" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.419910 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-dtgrm" podStartSLOduration=27.415064157 podStartE2EDuration="51.419885414s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.502031424 +0000 UTC m=+1268.144051984" lastFinishedPulling="2025-12-03 00:45:06.506852681 +0000 UTC m=+1292.148873241" observedRunningTime="2025-12-03 00:45:30.401586243 +0000 UTC m=+1316.043606803" watchObservedRunningTime="2025-12-03 00:45:30.419885414 +0000 UTC m=+1316.061905974" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.458622 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-g4n8g" podStartSLOduration=27.381746279 podStartE2EDuration="51.458583522s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.43010349 +0000 UTC m=+1268.072124050" lastFinishedPulling="2025-12-03 00:45:06.506940723 +0000 UTC m=+1292.148961293" observedRunningTime="2025-12-03 00:45:30.440759693 +0000 UTC m=+1316.082780253" watchObservedRunningTime="2025-12-03 00:45:30.458583522 +0000 UTC m=+1316.100604082" Dec 03 00:45:30 crc kubenswrapper[4912]: I1203 00:45:30.519712 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" podStartSLOduration=6.604722057 podStartE2EDuration="51.519671148s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.534618118 +0000 UTC m=+1268.176638678" lastFinishedPulling="2025-12-03 00:45:27.449567209 +0000 UTC m=+1313.091587769" observedRunningTime="2025-12-03 00:45:30.501883371 +0000 UTC m=+1316.143903951" watchObservedRunningTime="2025-12-03 00:45:30.519671148 +0000 UTC m=+1316.161691728" Dec 03 00:45:31 crc kubenswrapper[4912]: I1203 00:45:31.109337 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" event={"ID":"de517f42-de01-4dde-ac22-1a651a38926b","Type":"ContainerStarted","Data":"22363edb0b6ff84f2ac6d12999b1108b2269fba34f9e0789e211f9bf45332374"} Dec 03 00:45:31 crc kubenswrapper[4912]: I1203 00:45:31.110386 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" Dec 03 00:45:31 crc kubenswrapper[4912]: I1203 00:45:31.120335 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" event={"ID":"5118703e-2fdb-41d6-a76e-ab5d3f6adb95","Type":"ContainerStarted","Data":"353caabee6322d360ed5a79846cb70b8abae889a07564ff829028809bf4475b4"} Dec 03 00:45:31 crc kubenswrapper[4912]: E1203 00:45:31.127630 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" podUID="045ab366-9392-40f9-94f5-95c983f65176" Dec 03 00:45:31 crc kubenswrapper[4912]: E1203 00:45:31.128136 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:14cfad6ea2e7f7ecc4cb2aafceb9c61514b3d04b66668832d1e4ac3b19f1ab81\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" podUID="3e85fc81-3fbd-4af2-8c98-89892a4f3579" Dec 03 00:45:31 crc kubenswrapper[4912]: I1203 00:45:31.142127 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" podStartSLOduration=4.842415995 podStartE2EDuration="52.142103698s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.491633501 +0000 UTC m=+1268.133654061" lastFinishedPulling="2025-12-03 00:45:29.791321204 +0000 UTC m=+1315.433341764" observedRunningTime="2025-12-03 00:45:31.133994944 +0000 UTC m=+1316.776015504" watchObservedRunningTime="2025-12-03 00:45:31.142103698 +0000 UTC m=+1316.784124248" Dec 03 00:45:31 crc kubenswrapper[4912]: I1203 00:45:31.199501 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" podStartSLOduration=4.901542789 podStartE2EDuration="52.199462776s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.491828186 +0000 UTC m=+1268.133848746" lastFinishedPulling="2025-12-03 00:45:29.789748173 +0000 UTC m=+1315.431768733" observedRunningTime="2025-12-03 00:45:31.192418411 +0000 UTC m=+1316.834438961" watchObservedRunningTime="2025-12-03 00:45:31.199462776 +0000 UTC m=+1316.841483336" Dec 03 00:45:32 crc kubenswrapper[4912]: I1203 00:45:32.132660 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" Dec 03 00:45:32 crc kubenswrapper[4912]: I1203 00:45:32.133118 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-rjqjr" Dec 03 00:45:32 crc kubenswrapper[4912]: I1203 00:45:32.133410 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-7rstc" Dec 03 00:45:40 crc kubenswrapper[4912]: I1203 00:45:40.060390 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-tqphq" Dec 03 00:45:40 crc kubenswrapper[4912]: I1203 00:45:40.201896 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-jskq6" Dec 03 00:45:40 crc kubenswrapper[4912]: I1203 00:45:40.223858 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" event={"ID":"82b74254-de20-48a4-bfae-21afed7954cd","Type":"ContainerStarted","Data":"7571111547deb791ea85a78ebea90bad4881c39f2a5d46865ea6034bbfc30a92"} Dec 03 00:45:40 crc kubenswrapper[4912]: I1203 00:45:40.243651 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-rb4z7" Dec 03 00:45:40 crc kubenswrapper[4912]: I1203 00:45:40.636620 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-mjphw" Dec 03 00:45:40 crc kubenswrapper[4912]: I1203 00:45:40.695187 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-bdgdj" Dec 03 00:45:40 crc kubenswrapper[4912]: I1203 00:45:40.823200 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-vw7dk" Dec 03 00:45:41 crc kubenswrapper[4912]: I1203 00:45:41.238631 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" event={"ID":"82b74254-de20-48a4-bfae-21afed7954cd","Type":"ContainerStarted","Data":"20752ce7deee3d46b2209f7e5e73f345562486412e8d45d274d3493dba8fbb00"} Dec 03 00:45:41 crc kubenswrapper[4912]: I1203 00:45:41.238849 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" Dec 03 00:45:41 crc kubenswrapper[4912]: I1203 00:45:41.249578 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-8x84c" Dec 03 00:45:41 crc kubenswrapper[4912]: I1203 00:45:41.262847 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" podStartSLOduration=3.915364603 podStartE2EDuration="1m1.262828507s" podCreationTimestamp="2025-12-03 00:44:40 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.574882401 +0000 UTC m=+1268.216902961" lastFinishedPulling="2025-12-03 00:45:39.922346275 +0000 UTC m=+1325.564366865" observedRunningTime="2025-12-03 00:45:41.25909677 +0000 UTC m=+1326.901117340" watchObservedRunningTime="2025-12-03 00:45:41.262828507 +0000 UTC m=+1326.904849077" Dec 03 00:45:41 crc kubenswrapper[4912]: E1203 00:45:41.573760 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" podUID="b54a05e7-9cfd-401d-8e5d-87c8ed5ad321" Dec 03 00:45:43 crc kubenswrapper[4912]: I1203 00:45:43.449775 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-86464467f8-sq576" Dec 03 00:45:45 crc kubenswrapper[4912]: I1203 00:45:45.283276 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" event={"ID":"3e85fc81-3fbd-4af2-8c98-89892a4f3579","Type":"ContainerStarted","Data":"352dba0a3e49d885519f16a7a6105d3a16f722810032f93678968c278f690a64"} Dec 03 00:45:45 crc kubenswrapper[4912]: I1203 00:45:45.284404 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:45:45 crc kubenswrapper[4912]: I1203 00:45:45.287180 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" event={"ID":"045ab366-9392-40f9-94f5-95c983f65176","Type":"ContainerStarted","Data":"365fbd6cce4bfe60873d78c91c50517622cd5f3f6eccab55a43a4e37c8ff2a6a"} Dec 03 00:45:45 crc kubenswrapper[4912]: I1203 00:45:45.287569 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:45:45 crc kubenswrapper[4912]: I1203 00:45:45.321312 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" podStartSLOduration=28.276280538 podStartE2EDuration="1m5.32127975s" podCreationTimestamp="2025-12-03 00:44:40 +0000 UTC" firstStartedPulling="2025-12-03 00:45:07.250600803 +0000 UTC m=+1292.892621363" lastFinishedPulling="2025-12-03 00:45:44.295599975 +0000 UTC m=+1329.937620575" observedRunningTime="2025-12-03 00:45:45.313513695 +0000 UTC m=+1330.955534305" watchObservedRunningTime="2025-12-03 00:45:45.32127975 +0000 UTC m=+1330.963300320" Dec 03 00:45:45 crc kubenswrapper[4912]: I1203 00:45:45.337273 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" podStartSLOduration=29.248611134 podStartE2EDuration="1m6.337242389s" podCreationTimestamp="2025-12-03 00:44:39 +0000 UTC" firstStartedPulling="2025-12-03 00:45:07.937877462 +0000 UTC m=+1293.579898022" lastFinishedPulling="2025-12-03 00:45:45.026508717 +0000 UTC m=+1330.668529277" observedRunningTime="2025-12-03 00:45:45.332290179 +0000 UTC m=+1330.974310749" watchObservedRunningTime="2025-12-03 00:45:45.337242389 +0000 UTC m=+1330.979262959" Dec 03 00:45:48 crc kubenswrapper[4912]: I1203 00:45:48.078396 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:45:48 crc kubenswrapper[4912]: I1203 00:45:48.079927 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:45:51 crc kubenswrapper[4912]: I1203 00:45:51.210418 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5d66d5c957-xgv9t" Dec 03 00:45:54 crc kubenswrapper[4912]: I1203 00:45:54.386669 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" event={"ID":"b54a05e7-9cfd-401d-8e5d-87c8ed5ad321","Type":"ContainerStarted","Data":"e63240604d55ce5dfe1f496e21fefde6d7cb325d27bd3a316b7bec4f548abf9f"} Dec 03 00:45:54 crc kubenswrapper[4912]: I1203 00:45:54.420275 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-llhvt" podStartSLOduration=3.062293775 podStartE2EDuration="1m14.42024821s" podCreationTimestamp="2025-12-03 00:44:40 +0000 UTC" firstStartedPulling="2025-12-03 00:44:42.711058868 +0000 UTC m=+1268.353079418" lastFinishedPulling="2025-12-03 00:45:54.069013283 +0000 UTC m=+1339.711033853" observedRunningTime="2025-12-03 00:45:54.411799037 +0000 UTC m=+1340.053819637" watchObservedRunningTime="2025-12-03 00:45:54.42024821 +0000 UTC m=+1340.062268800" Dec 03 00:45:55 crc kubenswrapper[4912]: I1203 00:45:55.876009 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-lzff2" Dec 03 00:45:56 crc kubenswrapper[4912]: I1203 00:45:56.754401 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.948025 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tgxv2"] Dec 03 00:46:12 crc kubenswrapper[4912]: E1203 00:46:12.949250 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5be2251-7d9d-4360-8e88-d71bc8458cf8" containerName="collect-profiles" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.949264 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5be2251-7d9d-4360-8e88-d71bc8458cf8" containerName="collect-profiles" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.949494 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5be2251-7d9d-4360-8e88-d71bc8458cf8" containerName="collect-profiles" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.950410 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.952830 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.953224 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-b97nl" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.953715 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.954291 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.968174 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tgxv2"] Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.975370 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29l5g\" (UniqueName: \"kubernetes.io/projected/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-kube-api-access-29l5g\") pod \"dnsmasq-dns-675f4bcbfc-tgxv2\" (UID: \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:12 crc kubenswrapper[4912]: I1203 00:46:12.975457 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-config\") pod \"dnsmasq-dns-675f4bcbfc-tgxv2\" (UID: \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.020079 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vql96"] Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.027015 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.030116 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.048887 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vql96"] Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.076539 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-config\") pod \"dnsmasq-dns-78dd6ddcc-vql96\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.076617 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-config\") pod \"dnsmasq-dns-675f4bcbfc-tgxv2\" (UID: \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.076646 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7d6x\" (UniqueName: \"kubernetes.io/projected/a544c8a7-c831-4f82-bf4a-c329a423bb99-kube-api-access-t7d6x\") pod \"dnsmasq-dns-78dd6ddcc-vql96\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.076736 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vql96\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.076765 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29l5g\" (UniqueName: \"kubernetes.io/projected/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-kube-api-access-29l5g\") pod \"dnsmasq-dns-675f4bcbfc-tgxv2\" (UID: \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.078223 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-config\") pod \"dnsmasq-dns-675f4bcbfc-tgxv2\" (UID: \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.123346 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29l5g\" (UniqueName: \"kubernetes.io/projected/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-kube-api-access-29l5g\") pod \"dnsmasq-dns-675f4bcbfc-tgxv2\" (UID: \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.178470 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vql96\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.178563 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-config\") pod \"dnsmasq-dns-78dd6ddcc-vql96\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.178607 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7d6x\" (UniqueName: \"kubernetes.io/projected/a544c8a7-c831-4f82-bf4a-c329a423bb99-kube-api-access-t7d6x\") pod \"dnsmasq-dns-78dd6ddcc-vql96\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.179846 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-vql96\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.179872 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-config\") pod \"dnsmasq-dns-78dd6ddcc-vql96\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.197704 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7d6x\" (UniqueName: \"kubernetes.io/projected/a544c8a7-c831-4f82-bf4a-c329a423bb99-kube-api-access-t7d6x\") pod \"dnsmasq-dns-78dd6ddcc-vql96\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.267910 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.345652 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:13 crc kubenswrapper[4912]: I1203 00:46:13.837295 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tgxv2"] Dec 03 00:46:14 crc kubenswrapper[4912]: I1203 00:46:14.053145 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vql96"] Dec 03 00:46:14 crc kubenswrapper[4912]: W1203 00:46:14.059810 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda544c8a7_c831_4f82_bf4a_c329a423bb99.slice/crio-6101f1fcd4d69b5b20dddd8dbb612d2811a41bb17c66e015c560fae8618211c1 WatchSource:0}: Error finding container 6101f1fcd4d69b5b20dddd8dbb612d2811a41bb17c66e015c560fae8618211c1: Status 404 returned error can't find the container with id 6101f1fcd4d69b5b20dddd8dbb612d2811a41bb17c66e015c560fae8618211c1 Dec 03 00:46:14 crc kubenswrapper[4912]: I1203 00:46:14.617195 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" event={"ID":"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7","Type":"ContainerStarted","Data":"07afd68c3cd2b908ab23b573b8bb067f03bae9ca2421eedaf6f1a5975c90633a"} Dec 03 00:46:14 crc kubenswrapper[4912]: I1203 00:46:14.619195 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" event={"ID":"a544c8a7-c831-4f82-bf4a-c329a423bb99","Type":"ContainerStarted","Data":"6101f1fcd4d69b5b20dddd8dbb612d2811a41bb17c66e015c560fae8618211c1"} Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.052342 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tgxv2"] Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.067112 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q4drb"] Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.068944 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.098176 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q4drb"] Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.258985 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-config\") pod \"dnsmasq-dns-666b6646f7-q4drb\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.259044 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-q4drb\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.259095 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6n7p\" (UniqueName: \"kubernetes.io/projected/40178e9e-c6a8-4742-b5db-f090c3f462e1-kube-api-access-l6n7p\") pod \"dnsmasq-dns-666b6646f7-q4drb\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.360857 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-config\") pod \"dnsmasq-dns-666b6646f7-q4drb\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.360929 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-q4drb\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.360995 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6n7p\" (UniqueName: \"kubernetes.io/projected/40178e9e-c6a8-4742-b5db-f090c3f462e1-kube-api-access-l6n7p\") pod \"dnsmasq-dns-666b6646f7-q4drb\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.362328 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-q4drb\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.362868 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-config\") pod \"dnsmasq-dns-666b6646f7-q4drb\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.376037 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vql96"] Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.395352 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6n7p\" (UniqueName: \"kubernetes.io/projected/40178e9e-c6a8-4742-b5db-f090c3f462e1-kube-api-access-l6n7p\") pod \"dnsmasq-dns-666b6646f7-q4drb\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.407464 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.420408 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dt8zv"] Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.424087 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.443131 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dt8zv"] Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.566361 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dt8zv\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.566922 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t654h\" (UniqueName: \"kubernetes.io/projected/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-kube-api-access-t654h\") pod \"dnsmasq-dns-57d769cc4f-dt8zv\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.566997 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-config\") pod \"dnsmasq-dns-57d769cc4f-dt8zv\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.668898 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dt8zv\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.669988 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t654h\" (UniqueName: \"kubernetes.io/projected/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-kube-api-access-t654h\") pod \"dnsmasq-dns-57d769cc4f-dt8zv\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.670059 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-config\") pod \"dnsmasq-dns-57d769cc4f-dt8zv\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.669920 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-dt8zv\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.672642 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-config\") pod \"dnsmasq-dns-57d769cc4f-dt8zv\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.716873 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t654h\" (UniqueName: \"kubernetes.io/projected/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-kube-api-access-t654h\") pod \"dnsmasq-dns-57d769cc4f-dt8zv\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:16 crc kubenswrapper[4912]: I1203 00:46:16.760021 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.164301 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q4drb"] Dec 03 00:46:17 crc kubenswrapper[4912]: W1203 00:46:17.172405 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40178e9e_c6a8_4742_b5db_f090c3f462e1.slice/crio-e071d984f52c1a9333b0786f0ba7e20f7bef72a070f905815fa954b966d398c2 WatchSource:0}: Error finding container e071d984f52c1a9333b0786f0ba7e20f7bef72a070f905815fa954b966d398c2: Status 404 returned error can't find the container with id e071d984f52c1a9333b0786f0ba7e20f7bef72a070f905815fa954b966d398c2 Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.202035 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.203747 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.208553 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.208609 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.208862 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.208938 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-klxbd" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.208990 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.209081 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.209115 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.230777 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.382841 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.382976 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.383064 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/38bf4c96-49c7-432f-be55-159b75ed1243-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.383249 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/38bf4c96-49c7-432f-be55-159b75ed1243-pod-info\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.383604 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.383648 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.383913 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-config-data\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.384212 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.384563 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-server-conf\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.384892 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.385169 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhsx8\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-kube-api-access-mhsx8\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.464100 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dt8zv"] Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514372 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-config-data\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514693 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514724 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-server-conf\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514782 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514810 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhsx8\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-kube-api-access-mhsx8\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514861 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514883 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514905 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/38bf4c96-49c7-432f-be55-159b75ed1243-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514923 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/38bf4c96-49c7-432f-be55-159b75ed1243-pod-info\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514947 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.514968 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.517647 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.518739 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.520495 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-server-conf\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.521665 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.521750 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-config-data\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.522706 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.533771 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.533860 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.534507 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/38bf4c96-49c7-432f-be55-159b75ed1243-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.534773 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/38bf4c96-49c7-432f-be55-159b75ed1243-pod-info\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.571751 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.573579 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.579236 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.580372 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.581124 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.581391 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-q26vs" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.582204 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.582347 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.582198 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhsx8\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-kube-api-access-mhsx8\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.583228 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.584495 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.585791 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.682463 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-q4drb" event={"ID":"40178e9e-c6a8-4742-b5db-f090c3f462e1","Type":"ContainerStarted","Data":"e071d984f52c1a9333b0786f0ba7e20f7bef72a070f905815fa954b966d398c2"} Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.691164 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" event={"ID":"4cd22f06-6fc2-4edd-b0fc-44f4307878f0","Type":"ContainerStarted","Data":"eeb0ab237e8d2a20bc9f8e7166aa8afae8c54208108ffc3467f5f1d52545c43d"} Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.719680 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.719787 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.719828 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.719852 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.719887 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j6rf\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-kube-api-access-8j6rf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.719906 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.719926 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.719957 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.721501 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.721902 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.722147 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824520 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824609 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824628 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824666 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824686 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j6rf\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-kube-api-access-8j6rf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824704 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824730 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824749 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824802 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824847 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.824872 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.825827 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.831261 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.831641 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.834351 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.840448 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.844276 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.857080 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.857511 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.857618 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.858041 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.858151 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.865127 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j6rf\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-kube-api-access-8j6rf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.874689 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:17 crc kubenswrapper[4912]: I1203 00:46:17.994143 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.080673 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.080773 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.080900 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.082007 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d52c01534a739fd57a8ce1b61cd4e126b48cef8755e1aae547590b36373a992"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.082085 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://5d52c01534a739fd57a8ce1b61cd4e126b48cef8755e1aae547590b36373a992" gracePeriod=600 Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.668468 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.761619 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="5d52c01534a739fd57a8ce1b61cd4e126b48cef8755e1aae547590b36373a992" exitCode=0 Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.761692 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"5d52c01534a739fd57a8ce1b61cd4e126b48cef8755e1aae547590b36373a992"} Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.761733 4912 scope.go:117] "RemoveContainer" containerID="98a3aa8d06bd9638c9ea09b90a6f183e2b398dcf19a529ac1c64d4b6190f95cf" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.767823 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"38bf4c96-49c7-432f-be55-159b75ed1243","Type":"ContainerStarted","Data":"369dd507f45a672320ef8d9f416745f870e5836b37b05cb4d774b2cda8fcfa37"} Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.828606 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.839812 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.861253 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.862920 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.863060 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-kzfsl" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.864527 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.864666 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.872718 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.932555 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 00:46:18 crc kubenswrapper[4912]: W1203 00:46:18.964363 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef9c2da6_7cb1_48df_b51c_26ced3ec63c2.slice/crio-fbce2ecbf5f2597e41ee011d99013ea45de6413f6420e71f10f2d579b7519515 WatchSource:0}: Error finding container fbce2ecbf5f2597e41ee011d99013ea45de6413f6420e71f10f2d579b7519515: Status 404 returned error can't find the container with id fbce2ecbf5f2597e41ee011d99013ea45de6413f6420e71f10f2d579b7519515 Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.993464 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a1a55bf-bc0f-4004-b540-ead98e30d25a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.993512 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1a55bf-bc0f-4004-b540-ead98e30d25a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.993540 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a1a55bf-bc0f-4004-b540-ead98e30d25a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.993589 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a1a55bf-bc0f-4004-b540-ead98e30d25a-kolla-config\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.993614 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f25c8\" (UniqueName: \"kubernetes.io/projected/9a1a55bf-bc0f-4004-b540-ead98e30d25a-kube-api-access-f25c8\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.993729 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.993881 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a1a55bf-bc0f-4004-b540-ead98e30d25a-config-data-default\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:18 crc kubenswrapper[4912]: I1203 00:46:18.993919 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a1a55bf-bc0f-4004-b540-ead98e30d25a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.096980 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.097072 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a1a55bf-bc0f-4004-b540-ead98e30d25a-config-data-default\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.097102 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a1a55bf-bc0f-4004-b540-ead98e30d25a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.097179 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a1a55bf-bc0f-4004-b540-ead98e30d25a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.097210 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1a55bf-bc0f-4004-b540-ead98e30d25a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.097228 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a1a55bf-bc0f-4004-b540-ead98e30d25a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.097262 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a1a55bf-bc0f-4004-b540-ead98e30d25a-kolla-config\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.097304 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f25c8\" (UniqueName: \"kubernetes.io/projected/9a1a55bf-bc0f-4004-b540-ead98e30d25a-kube-api-access-f25c8\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.098242 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9a1a55bf-bc0f-4004-b540-ead98e30d25a-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.098674 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.098701 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9a1a55bf-bc0f-4004-b540-ead98e30d25a-kolla-config\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.100212 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a1a55bf-bc0f-4004-b540-ead98e30d25a-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.102020 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9a1a55bf-bc0f-4004-b540-ead98e30d25a-config-data-default\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.107925 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a1a55bf-bc0f-4004-b540-ead98e30d25a-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.118791 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a1a55bf-bc0f-4004-b540-ead98e30d25a-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.127589 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f25c8\" (UniqueName: \"kubernetes.io/projected/9a1a55bf-bc0f-4004-b540-ead98e30d25a-kube-api-access-f25c8\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.147243 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"9a1a55bf-bc0f-4004-b540-ead98e30d25a\") " pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.462969 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 00:46:19 crc kubenswrapper[4912]: I1203 00:46:19.811382 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2","Type":"ContainerStarted","Data":"fbce2ecbf5f2597e41ee011d99013ea45de6413f6420e71f10f2d579b7519515"} Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.076163 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.118102 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.144506 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.168391 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xt6gt" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.169601 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.169778 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.172106 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.199653 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.356015 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e835f70a-04f2-43d3-ba92-6a6d20216a7d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.356115 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e835f70a-04f2-43d3-ba92-6a6d20216a7d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.356148 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e835f70a-04f2-43d3-ba92-6a6d20216a7d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.356212 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddssh\" (UniqueName: \"kubernetes.io/projected/e835f70a-04f2-43d3-ba92-6a6d20216a7d-kube-api-access-ddssh\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.356322 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e835f70a-04f2-43d3-ba92-6a6d20216a7d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.356357 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.356392 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e835f70a-04f2-43d3-ba92-6a6d20216a7d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.356420 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e835f70a-04f2-43d3-ba92-6a6d20216a7d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.404752 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.406153 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.418659 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.419285 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.420092 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.420937 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-54xx6" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.458015 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e835f70a-04f2-43d3-ba92-6a6d20216a7d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.458107 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddssh\" (UniqueName: \"kubernetes.io/projected/e835f70a-04f2-43d3-ba92-6a6d20216a7d-kube-api-access-ddssh\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.458153 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e835f70a-04f2-43d3-ba92-6a6d20216a7d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.458216 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.458250 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e835f70a-04f2-43d3-ba92-6a6d20216a7d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.458280 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e835f70a-04f2-43d3-ba92-6a6d20216a7d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.458341 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e835f70a-04f2-43d3-ba92-6a6d20216a7d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.458376 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e835f70a-04f2-43d3-ba92-6a6d20216a7d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.460093 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.462410 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e835f70a-04f2-43d3-ba92-6a6d20216a7d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.467456 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e835f70a-04f2-43d3-ba92-6a6d20216a7d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.468065 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e835f70a-04f2-43d3-ba92-6a6d20216a7d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.468563 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e835f70a-04f2-43d3-ba92-6a6d20216a7d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.469102 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e835f70a-04f2-43d3-ba92-6a6d20216a7d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.477521 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e835f70a-04f2-43d3-ba92-6a6d20216a7d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.487929 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddssh\" (UniqueName: \"kubernetes.io/projected/e835f70a-04f2-43d3-ba92-6a6d20216a7d-kube-api-access-ddssh\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.526131 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e835f70a-04f2-43d3-ba92-6a6d20216a7d\") " pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.551641 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.561331 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e909a03f-b885-46d4-a139-833410cf9ff1-config-data\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.561520 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e909a03f-b885-46d4-a139-833410cf9ff1-kolla-config\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.561581 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e909a03f-b885-46d4-a139-833410cf9ff1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.561842 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e909a03f-b885-46d4-a139-833410cf9ff1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.561940 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6j2z\" (UniqueName: \"kubernetes.io/projected/e909a03f-b885-46d4-a139-833410cf9ff1-kube-api-access-l6j2z\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.663628 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e909a03f-b885-46d4-a139-833410cf9ff1-kolla-config\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.663796 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e909a03f-b885-46d4-a139-833410cf9ff1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.663852 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e909a03f-b885-46d4-a139-833410cf9ff1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.663911 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6j2z\" (UniqueName: \"kubernetes.io/projected/e909a03f-b885-46d4-a139-833410cf9ff1-kube-api-access-l6j2z\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.664006 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e909a03f-b885-46d4-a139-833410cf9ff1-config-data\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.665230 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e909a03f-b885-46d4-a139-833410cf9ff1-kolla-config\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.665700 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e909a03f-b885-46d4-a139-833410cf9ff1-config-data\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.692898 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e909a03f-b885-46d4-a139-833410cf9ff1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.700316 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/e909a03f-b885-46d4-a139-833410cf9ff1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.779650 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6j2z\" (UniqueName: \"kubernetes.io/projected/e909a03f-b885-46d4-a139-833410cf9ff1-kube-api-access-l6j2z\") pod \"memcached-0\" (UID: \"e909a03f-b885-46d4-a139-833410cf9ff1\") " pod="openstack/memcached-0" Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.963820 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a"} Dec 03 00:46:20 crc kubenswrapper[4912]: I1203 00:46:20.974627 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9a1a55bf-bc0f-4004-b540-ead98e30d25a","Type":"ContainerStarted","Data":"5fc2dba92f57fe50c1d62ee81adc2bfa857a403c4f8474fe9da44468f7d93c06"} Dec 03 00:46:21 crc kubenswrapper[4912]: I1203 00:46:21.051223 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 00:46:21 crc kubenswrapper[4912]: I1203 00:46:21.566366 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 00:46:21 crc kubenswrapper[4912]: I1203 00:46:21.791581 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 00:46:21 crc kubenswrapper[4912]: W1203 00:46:21.810599 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode909a03f_b885_46d4_a139_833410cf9ff1.slice/crio-3e49a458f5c7a9467ffbd093ef4b58237dba4b6f8e94092e6a2b766bb69a12cf WatchSource:0}: Error finding container 3e49a458f5c7a9467ffbd093ef4b58237dba4b6f8e94092e6a2b766bb69a12cf: Status 404 returned error can't find the container with id 3e49a458f5c7a9467ffbd093ef4b58237dba4b6f8e94092e6a2b766bb69a12cf Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.057333 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e909a03f-b885-46d4-a139-833410cf9ff1","Type":"ContainerStarted","Data":"3e49a458f5c7a9467ffbd093ef4b58237dba4b6f8e94092e6a2b766bb69a12cf"} Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.068060 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e835f70a-04f2-43d3-ba92-6a6d20216a7d","Type":"ContainerStarted","Data":"35728e42a4a6fd839325e55cfa1560000af3a7fc7bfc868f805f5778c41d6c0e"} Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.561256 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.574759 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.587634 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-x5jjw" Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.627576 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcntt\" (UniqueName: \"kubernetes.io/projected/d4611c7f-5acd-4fb6-a984-ef56f9feeb72-kube-api-access-hcntt\") pod \"kube-state-metrics-0\" (UID: \"d4611c7f-5acd-4fb6-a984-ef56f9feeb72\") " pod="openstack/kube-state-metrics-0" Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.627633 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.735397 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcntt\" (UniqueName: \"kubernetes.io/projected/d4611c7f-5acd-4fb6-a984-ef56f9feeb72-kube-api-access-hcntt\") pod \"kube-state-metrics-0\" (UID: \"d4611c7f-5acd-4fb6-a984-ef56f9feeb72\") " pod="openstack/kube-state-metrics-0" Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.787316 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcntt\" (UniqueName: \"kubernetes.io/projected/d4611c7f-5acd-4fb6-a984-ef56f9feeb72-kube-api-access-hcntt\") pod \"kube-state-metrics-0\" (UID: \"d4611c7f-5acd-4fb6-a984-ef56f9feeb72\") " pod="openstack/kube-state-metrics-0" Dec 03 00:46:22 crc kubenswrapper[4912]: I1203 00:46:22.921955 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.394676 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5"] Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.397406 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.405162 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-9hdl8" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.444685 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.450398 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5"] Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.480924 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkl58\" (UniqueName: \"kubernetes.io/projected/9d7ffd59-b442-43e2-b7b4-5a2fc42507d3-kube-api-access-nkl58\") pod \"observability-ui-dashboards-7d5fb4cbfb-96rq5\" (UID: \"9d7ffd59-b442-43e2-b7b4-5a2fc42507d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.481267 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7ffd59-b442-43e2-b7b4-5a2fc42507d3-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-96rq5\" (UID: \"9d7ffd59-b442-43e2-b7b4-5a2fc42507d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.583439 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkl58\" (UniqueName: \"kubernetes.io/projected/9d7ffd59-b442-43e2-b7b4-5a2fc42507d3-kube-api-access-nkl58\") pod \"observability-ui-dashboards-7d5fb4cbfb-96rq5\" (UID: \"9d7ffd59-b442-43e2-b7b4-5a2fc42507d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.583554 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7ffd59-b442-43e2-b7b4-5a2fc42507d3-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-96rq5\" (UID: \"9d7ffd59-b442-43e2-b7b4-5a2fc42507d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.616415 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d7ffd59-b442-43e2-b7b4-5a2fc42507d3-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-96rq5\" (UID: \"9d7ffd59-b442-43e2-b7b4-5a2fc42507d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.656265 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkl58\" (UniqueName: \"kubernetes.io/projected/9d7ffd59-b442-43e2-b7b4-5a2fc42507d3-kube-api-access-nkl58\") pod \"observability-ui-dashboards-7d5fb4cbfb-96rq5\" (UID: \"9d7ffd59-b442-43e2-b7b4-5a2fc42507d3\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.778920 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.826362 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.830067 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.835515 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.838467 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.839818 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-w4fkg" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.841048 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.843099 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.843390 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.885561 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.996938 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.997005 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.997125 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.997229 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.997263 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.997286 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.997310 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57brj\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-kube-api-access-57brj\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:23 crc kubenswrapper[4912]: I1203 00:46:23.997348 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.024947 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-bbcd684f4-hp2d8"] Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.026408 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.046957 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-bbcd684f4-hp2d8"] Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.106810 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-service-ca\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.114141 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.114288 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.114400 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.114515 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57brj\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-kube-api-access-57brj\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.114616 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-console-config\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.114731 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.114893 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12373aea-a902-4c10-a5f0-d2bfa255dc55-console-serving-cert\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.115086 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r4gg\" (UniqueName: \"kubernetes.io/projected/12373aea-a902-4c10-a5f0-d2bfa255dc55-kube-api-access-5r4gg\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.115210 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-oauth-serving-cert\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.115319 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12373aea-a902-4c10-a5f0-d2bfa255dc55-console-oauth-config\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.115457 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-trusted-ca-bundle\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.115602 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.115688 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.115762 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.115789 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.118078 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.158743 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.159382 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.159866 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.160320 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.184823 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.210757 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57brj\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-kube-api-access-57brj\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.240383 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-service-ca\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.240565 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-console-config\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.240700 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12373aea-a902-4c10-a5f0-d2bfa255dc55-console-serving-cert\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.240803 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r4gg\" (UniqueName: \"kubernetes.io/projected/12373aea-a902-4c10-a5f0-d2bfa255dc55-kube-api-access-5r4gg\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.240876 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-oauth-serving-cert\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.240923 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12373aea-a902-4c10-a5f0-d2bfa255dc55-console-oauth-config\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.240964 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-trusted-ca-bundle\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.241688 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-service-ca\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.242033 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-console-config\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.251460 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-oauth-serving-cert\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.257290 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/12373aea-a902-4c10-a5f0-d2bfa255dc55-console-oauth-config\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.259575 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12373aea-a902-4c10-a5f0-d2bfa255dc55-trusted-ca-bundle\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.265454 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/12373aea-a902-4c10-a5f0-d2bfa255dc55-console-serving-cert\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.291924 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.293406 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r4gg\" (UniqueName: \"kubernetes.io/projected/12373aea-a902-4c10-a5f0-d2bfa255dc55-kube-api-access-5r4gg\") pod \"console-bbcd684f4-hp2d8\" (UID: \"12373aea-a902-4c10-a5f0-d2bfa255dc55\") " pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.387322 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:46:24 crc kubenswrapper[4912]: I1203 00:46:24.478302 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.882048 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.889484 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.894823 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.895004 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-2gpvb" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.895141 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.895474 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.895761 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.915024 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.941170 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.941257 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.941289 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.941313 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.941462 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxs84\" (UniqueName: \"kubernetes.io/projected/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-kube-api-access-pxs84\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.941580 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-config\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.941617 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:26 crc kubenswrapper[4912]: I1203 00:46:26.941680 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.044421 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.044504 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.044620 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxs84\" (UniqueName: \"kubernetes.io/projected/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-kube-api-access-pxs84\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.044667 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-config\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.044696 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.044725 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.044756 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.044784 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.046537 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-config\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.047382 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.048150 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.049599 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.054714 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.056755 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gbbh7"] Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.058531 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.062168 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-6xrnb" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.062827 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.063945 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.064214 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.064854 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.099745 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6lm7j"] Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.099775 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxs84\" (UniqueName: \"kubernetes.io/projected/ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026-kube-api-access-pxs84\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.109466 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.113928 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gbbh7"] Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.126298 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6lm7j"] Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.127603 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026\") " pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.154912 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-combined-ca-bundle\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.154995 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-var-log\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155025 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-var-run-ovn\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155092 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-etc-ovs\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155142 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-var-log-ovn\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155173 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-scripts\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155198 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-ovn-controller-tls-certs\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155226 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsfkk\" (UniqueName: \"kubernetes.io/projected/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-kube-api-access-bsfkk\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155248 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-var-lib\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155293 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btvfn\" (UniqueName: \"kubernetes.io/projected/2c613cf4-0585-466f-8db2-4c4f4b4765f0-kube-api-access-btvfn\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155315 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c613cf4-0585-466f-8db2-4c4f4b4765f0-scripts\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155351 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-var-run\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.155373 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-var-run\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.227092 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.259071 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-var-run\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.259378 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-var-run\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.259534 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-combined-ca-bundle\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.259650 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-var-log\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.259749 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-var-run-ovn\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.259863 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-etc-ovs\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.259977 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-var-log-ovn\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.260079 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-scripts\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.260158 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-ovn-controller-tls-certs\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.260259 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsfkk\" (UniqueName: \"kubernetes.io/projected/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-kube-api-access-bsfkk\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.260344 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-var-lib\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.260501 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btvfn\" (UniqueName: \"kubernetes.io/projected/2c613cf4-0585-466f-8db2-4c4f4b4765f0-kube-api-access-btvfn\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.260581 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c613cf4-0585-466f-8db2-4c4f4b4765f0-scripts\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.261353 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-var-run\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.263870 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-var-log-ovn\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.263931 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-var-run\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.263977 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c613cf4-0585-466f-8db2-4c4f4b4765f0-scripts\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.264292 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-var-lib\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.264739 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-var-run-ovn\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.264912 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-etc-ovs\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.264927 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2c613cf4-0585-466f-8db2-4c4f4b4765f0-var-log\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.288264 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-scripts\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.293837 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-ovn-controller-tls-certs\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.293995 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-combined-ca-bundle\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.294780 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsfkk\" (UniqueName: \"kubernetes.io/projected/dd4efef6-5ce3-444a-a464-1f18f7cc2db4-kube-api-access-bsfkk\") pod \"ovn-controller-gbbh7\" (UID: \"dd4efef6-5ce3-444a-a464-1f18f7cc2db4\") " pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.329539 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btvfn\" (UniqueName: \"kubernetes.io/projected/2c613cf4-0585-466f-8db2-4c4f4b4765f0-kube-api-access-btvfn\") pod \"ovn-controller-ovs-6lm7j\" (UID: \"2c613cf4-0585-466f-8db2-4c4f4b4765f0\") " pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.537515 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gbbh7" Dec 03 00:46:27 crc kubenswrapper[4912]: I1203 00:46:27.563148 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.302553 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.306979 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.311640 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-x284r" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.321634 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.321775 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.321829 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.355945 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.463770 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0dca179-6a1f-41ad-bad5-1d193e2583c8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.463828 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7qvz\" (UniqueName: \"kubernetes.io/projected/d0dca179-6a1f-41ad-bad5-1d193e2583c8-kube-api-access-f7qvz\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.463868 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0dca179-6a1f-41ad-bad5-1d193e2583c8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.463945 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0dca179-6a1f-41ad-bad5-1d193e2583c8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.463973 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d0dca179-6a1f-41ad-bad5-1d193e2583c8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.464084 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0dca179-6a1f-41ad-bad5-1d193e2583c8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.464111 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0dca179-6a1f-41ad-bad5-1d193e2583c8-config\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.464139 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.565986 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0dca179-6a1f-41ad-bad5-1d193e2583c8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.566048 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7qvz\" (UniqueName: \"kubernetes.io/projected/d0dca179-6a1f-41ad-bad5-1d193e2583c8-kube-api-access-f7qvz\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.566091 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0dca179-6a1f-41ad-bad5-1d193e2583c8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.566117 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0dca179-6a1f-41ad-bad5-1d193e2583c8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.566136 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d0dca179-6a1f-41ad-bad5-1d193e2583c8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.566173 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0dca179-6a1f-41ad-bad5-1d193e2583c8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.566201 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0dca179-6a1f-41ad-bad5-1d193e2583c8-config\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.566224 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.566762 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.568837 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d0dca179-6a1f-41ad-bad5-1d193e2583c8-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.569826 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0dca179-6a1f-41ad-bad5-1d193e2583c8-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.570459 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0dca179-6a1f-41ad-bad5-1d193e2583c8-config\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.575871 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0dca179-6a1f-41ad-bad5-1d193e2583c8-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.581142 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0dca179-6a1f-41ad-bad5-1d193e2583c8-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.586020 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d0dca179-6a1f-41ad-bad5-1d193e2583c8-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.596035 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7qvz\" (UniqueName: \"kubernetes.io/projected/d0dca179-6a1f-41ad-bad5-1d193e2583c8-kube-api-access-f7qvz\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.604629 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"d0dca179-6a1f-41ad-bad5-1d193e2583c8\") " pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:30 crc kubenswrapper[4912]: I1203 00:46:30.640534 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.112226 4912 scope.go:117] "RemoveContainer" containerID="91e8e0b6c6ed4d3a700474a6ac09ebf880e314ae33c9f31a393a47e1eaf5b870" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.445976 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4smn9"] Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.448288 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.460042 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4smn9"] Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.594986 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-utilities\") pod \"redhat-operators-4smn9\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.595085 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdzv5\" (UniqueName: \"kubernetes.io/projected/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-kube-api-access-jdzv5\") pod \"redhat-operators-4smn9\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.595118 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-catalog-content\") pod \"redhat-operators-4smn9\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.698713 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-utilities\") pod \"redhat-operators-4smn9\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.698800 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdzv5\" (UniqueName: \"kubernetes.io/projected/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-kube-api-access-jdzv5\") pod \"redhat-operators-4smn9\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.698833 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-catalog-content\") pod \"redhat-operators-4smn9\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.699405 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-catalog-content\") pod \"redhat-operators-4smn9\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.699515 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-utilities\") pod \"redhat-operators-4smn9\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.745328 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdzv5\" (UniqueName: \"kubernetes.io/projected/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-kube-api-access-jdzv5\") pod \"redhat-operators-4smn9\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:35 crc kubenswrapper[4912]: I1203 00:46:35.778807 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:46:39 crc kubenswrapper[4912]: I1203 00:46:39.919751 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gbbh7"] Dec 03 00:46:56 crc kubenswrapper[4912]: E1203 00:46:56.030553 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 03 00:46:56 crc kubenswrapper[4912]: E1203 00:46:56.031375 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ddssh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(e835f70a-04f2-43d3-ba92-6a6d20216a7d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:46:56 crc kubenswrapper[4912]: E1203 00:46:56.033695 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="e835f70a-04f2-43d3-ba92-6a6d20216a7d" Dec 03 00:46:56 crc kubenswrapper[4912]: W1203 00:46:56.094426 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd4efef6_5ce3_444a_a464_1f18f7cc2db4.slice/crio-fce2fd4e39d6da14bc8cf66401a9dcb53515c74bb452cb0c09b2496a13fd9683 WatchSource:0}: Error finding container fce2fd4e39d6da14bc8cf66401a9dcb53515c74bb452cb0c09b2496a13fd9683: Status 404 returned error can't find the container with id fce2fd4e39d6da14bc8cf66401a9dcb53515c74bb452cb0c09b2496a13fd9683 Dec 03 00:46:56 crc kubenswrapper[4912]: I1203 00:46:56.119472 4912 scope.go:117] "RemoveContainer" containerID="1723885e60187b448ad9bfc53b0dde1b1f0d18932c0edd234db8f48ec1e90bab" Dec 03 00:46:56 crc kubenswrapper[4912]: I1203 00:46:56.210962 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:46:56 crc kubenswrapper[4912]: I1203 00:46:56.583522 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 00:46:56 crc kubenswrapper[4912]: I1203 00:46:56.590495 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gbbh7" event={"ID":"dd4efef6-5ce3-444a-a464-1f18f7cc2db4","Type":"ContainerStarted","Data":"fce2fd4e39d6da14bc8cf66401a9dcb53515c74bb452cb0c09b2496a13fd9683"} Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.198162 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.198348 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l6n7p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-q4drb_openstack(40178e9e-c6a8-4742-b5db-f090c3f462e1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.199978 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-q4drb" podUID="40178e9e-c6a8-4742-b5db-f090c3f462e1" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.219227 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.219590 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t654h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-dt8zv_openstack(4cd22f06-6fc2-4edd-b0fc-44f4307878f0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.220715 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.220761 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" podUID="4cd22f06-6fc2-4edd-b0fc-44f4307878f0" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.220866 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t7d6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-vql96_openstack(a544c8a7-c831-4f82-bf4a-c329a423bb99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.222167 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" podUID="a544c8a7-c831-4f82-bf4a-c329a423bb99" Dec 03 00:46:57 crc kubenswrapper[4912]: W1203 00:46:57.229417 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4611c7f_5acd_4fb6_a984_ef56f9feeb72.slice/crio-8cf0489ab34dad7a912e34bf697b4521a5bc5481a8db469df0f5c91ecdee2c89 WatchSource:0}: Error finding container 8cf0489ab34dad7a912e34bf697b4521a5bc5481a8db469df0f5c91ecdee2c89: Status 404 returned error can't find the container with id 8cf0489ab34dad7a912e34bf697b4521a5bc5481a8db469df0f5c91ecdee2c89 Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.244254 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.244627 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-29l5g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-tgxv2_openstack(cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.245970 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" podUID="cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7" Dec 03 00:46:57 crc kubenswrapper[4912]: I1203 00:46:57.258289 4912 scope.go:117] "RemoveContainer" containerID="0de3bc9625cbd655479d8e93a4bcd8acad44b535c43ed94a71bcf124fe5d773f" Dec 03 00:46:57 crc kubenswrapper[4912]: I1203 00:46:57.602594 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d4611c7f-5acd-4fb6-a984-ef56f9feeb72","Type":"ContainerStarted","Data":"8cf0489ab34dad7a912e34bf697b4521a5bc5481a8db469df0f5c91ecdee2c89"} Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.611754 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-q4drb" podUID="40178e9e-c6a8-4742-b5db-f090c3f462e1" Dec 03 00:46:57 crc kubenswrapper[4912]: E1203 00:46:57.611865 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" podUID="4cd22f06-6fc2-4edd-b0fc-44f4307878f0" Dec 03 00:46:57 crc kubenswrapper[4912]: I1203 00:46:57.804633 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5"] Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.145045 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.323992 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4smn9"] Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.342689 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.352318 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-bbcd684f4-hp2d8"] Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.461324 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6lm7j"] Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.620804 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" event={"ID":"9d7ffd59-b442-43e2-b7b4-5a2fc42507d3","Type":"ContainerStarted","Data":"3420815eae64ded0ecfecc631104beac29e1015c0801215f1b377b136dc920e6"} Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.624186 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" event={"ID":"a544c8a7-c831-4f82-bf4a-c329a423bb99","Type":"ContainerDied","Data":"6101f1fcd4d69b5b20dddd8dbb612d2811a41bb17c66e015c560fae8618211c1"} Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.624221 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6101f1fcd4d69b5b20dddd8dbb612d2811a41bb17c66e015c560fae8618211c1" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.641618 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9a1a55bf-bc0f-4004-b540-ead98e30d25a","Type":"ContainerStarted","Data":"c950221ced448322cba921a855d78d77ecfaef014bc38c222adfa0215c9becd7"} Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.650602 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e909a03f-b885-46d4-a139-833410cf9ff1","Type":"ContainerStarted","Data":"81ed8b89ff2429d91af794c040cf53fdfe4f409e1767ae27dbd6bb0b5b62a436"} Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.651902 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.678634 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d0dca179-6a1f-41ad-bad5-1d193e2583c8","Type":"ContainerStarted","Data":"cddcbe6094cf61254e528847cc30f82c126008e32fac819d7053b818879eaef6"} Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.687681 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e835f70a-04f2-43d3-ba92-6a6d20216a7d","Type":"ContainerStarted","Data":"e4acef7733793b833a44fb12bfd35cdae9e679940a77dd7e63ac3db9a6a7bb95"} Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.692511 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" event={"ID":"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7","Type":"ContainerDied","Data":"07afd68c3cd2b908ab23b573b8bb067f03bae9ca2421eedaf6f1a5975c90633a"} Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.692583 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07afd68c3cd2b908ab23b573b8bb067f03bae9ca2421eedaf6f1a5975c90633a" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.716451 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.717081 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.276735945 podStartE2EDuration="38.717070447s" podCreationTimestamp="2025-12-03 00:46:20 +0000 UTC" firstStartedPulling="2025-12-03 00:46:21.81790617 +0000 UTC m=+1367.459926730" lastFinishedPulling="2025-12-03 00:46:57.258240672 +0000 UTC m=+1402.900261232" observedRunningTime="2025-12-03 00:46:58.705931884 +0000 UTC m=+1404.347952444" watchObservedRunningTime="2025-12-03 00:46:58.717070447 +0000 UTC m=+1404.359091007" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.725812 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.836046 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-config\") pod \"a544c8a7-c831-4f82-bf4a-c329a423bb99\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.836426 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-dns-svc\") pod \"a544c8a7-c831-4f82-bf4a-c329a423bb99\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.836562 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29l5g\" (UniqueName: \"kubernetes.io/projected/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-kube-api-access-29l5g\") pod \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\" (UID: \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\") " Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.836639 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-config" (OuterVolumeSpecName: "config") pod "a544c8a7-c831-4f82-bf4a-c329a423bb99" (UID: "a544c8a7-c831-4f82-bf4a-c329a423bb99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.836733 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-config\") pod \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\" (UID: \"cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7\") " Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.836788 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7d6x\" (UniqueName: \"kubernetes.io/projected/a544c8a7-c831-4f82-bf4a-c329a423bb99-kube-api-access-t7d6x\") pod \"a544c8a7-c831-4f82-bf4a-c329a423bb99\" (UID: \"a544c8a7-c831-4f82-bf4a-c329a423bb99\") " Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.836842 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a544c8a7-c831-4f82-bf4a-c329a423bb99" (UID: "a544c8a7-c831-4f82-bf4a-c329a423bb99"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.837291 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.837313 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a544c8a7-c831-4f82-bf4a-c329a423bb99-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.838147 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-config" (OuterVolumeSpecName: "config") pod "cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7" (UID: "cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.940150 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.990127 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-kube-api-access-29l5g" (OuterVolumeSpecName: "kube-api-access-29l5g") pod "cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7" (UID: "cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7"). InnerVolumeSpecName "kube-api-access-29l5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:46:58 crc kubenswrapper[4912]: I1203 00:46:58.990893 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a544c8a7-c831-4f82-bf4a-c329a423bb99-kube-api-access-t7d6x" (OuterVolumeSpecName: "kube-api-access-t7d6x") pod "a544c8a7-c831-4f82-bf4a-c329a423bb99" (UID: "a544c8a7-c831-4f82-bf4a-c329a423bb99"). InnerVolumeSpecName "kube-api-access-t7d6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.044771 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7d6x\" (UniqueName: \"kubernetes.io/projected/a544c8a7-c831-4f82-bf4a-c329a423bb99-kube-api-access-t7d6x\") on node \"crc\" DevicePath \"\"" Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.044816 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29l5g\" (UniqueName: \"kubernetes.io/projected/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7-kube-api-access-29l5g\") on node \"crc\" DevicePath \"\"" Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.134549 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 00:46:59 crc kubenswrapper[4912]: W1203 00:46:59.375748 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad3ecd9c_b9b1_41c9_bbc9_ef5674f9e026.slice/crio-f0629fa0f03a0be32d3d67e9412a6b72940c375df2436f01b6c525cf918f0683 WatchSource:0}: Error finding container f0629fa0f03a0be32d3d67e9412a6b72940c375df2436f01b6c525cf918f0683: Status 404 returned error can't find the container with id f0629fa0f03a0be32d3d67e9412a6b72940c375df2436f01b6c525cf918f0683 Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.703401 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4smn9" event={"ID":"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f","Type":"ContainerStarted","Data":"d28a8329f9eb3f64a1e9c57aaac6b84d983182e4f548711e18f84f57d919e1ea"} Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.705051 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026","Type":"ContainerStarted","Data":"f0629fa0f03a0be32d3d67e9412a6b72940c375df2436f01b6c525cf918f0683"} Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.706657 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6lm7j" event={"ID":"2c613cf4-0585-466f-8db2-4c4f4b4765f0","Type":"ContainerStarted","Data":"2582870d28df0d27709976cbba059826b5a998a6c12f65bbd781ff7f38e273a2"} Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.709678 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerStarted","Data":"80dc7755a28ec37661ffcee2c27efedc08f43e0c2c0241515e53bbd765a9dccb"} Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.712138 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2","Type":"ContainerStarted","Data":"d601bbaa0a1e7902f9bd53059640462774984a5b945e216fc8163f7c3ae23a3c"} Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.713834 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-bbcd684f4-hp2d8" event={"ID":"12373aea-a902-4c10-a5f0-d2bfa255dc55","Type":"ContainerStarted","Data":"3f57443887f75b13fb128ec4ab485a17f560b8f9ccae7028541c4320a09eca41"} Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.718468 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-vql96" Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.718420 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"38bf4c96-49c7-432f-be55-159b75ed1243","Type":"ContainerStarted","Data":"dff02ed2edeeec80c4c16589ce57d80d6f4bff8f20597991b6274a1f56ab981a"} Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.718588 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tgxv2" Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.813651 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vql96"] Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.824022 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-vql96"] Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.849265 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tgxv2"] Dec 03 00:46:59 crc kubenswrapper[4912]: I1203 00:46:59.856216 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tgxv2"] Dec 03 00:47:00 crc kubenswrapper[4912]: I1203 00:47:00.610021 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a544c8a7-c831-4f82-bf4a-c329a423bb99" path="/var/lib/kubelet/pods/a544c8a7-c831-4f82-bf4a-c329a423bb99/volumes" Dec 03 00:47:00 crc kubenswrapper[4912]: I1203 00:47:00.610902 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7" path="/var/lib/kubelet/pods/cb9b0edb-f2dd-4aa5-8e18-e792109aa6a7/volumes" Dec 03 00:47:01 crc kubenswrapper[4912]: I1203 00:47:01.765954 4912 generic.go:334] "Generic (PLEG): container finished" podID="9a1a55bf-bc0f-4004-b540-ead98e30d25a" containerID="c950221ced448322cba921a855d78d77ecfaef014bc38c222adfa0215c9becd7" exitCode=0 Dec 03 00:47:01 crc kubenswrapper[4912]: I1203 00:47:01.766059 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9a1a55bf-bc0f-4004-b540-ead98e30d25a","Type":"ContainerDied","Data":"c950221ced448322cba921a855d78d77ecfaef014bc38c222adfa0215c9becd7"} Dec 03 00:47:01 crc kubenswrapper[4912]: I1203 00:47:01.771815 4912 generic.go:334] "Generic (PLEG): container finished" podID="e835f70a-04f2-43d3-ba92-6a6d20216a7d" containerID="e4acef7733793b833a44fb12bfd35cdae9e679940a77dd7e63ac3db9a6a7bb95" exitCode=0 Dec 03 00:47:01 crc kubenswrapper[4912]: I1203 00:47:01.772124 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e835f70a-04f2-43d3-ba92-6a6d20216a7d","Type":"ContainerDied","Data":"e4acef7733793b833a44fb12bfd35cdae9e679940a77dd7e63ac3db9a6a7bb95"} Dec 03 00:47:01 crc kubenswrapper[4912]: I1203 00:47:01.778491 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-bbcd684f4-hp2d8" event={"ID":"12373aea-a902-4c10-a5f0-d2bfa255dc55","Type":"ContainerStarted","Data":"6c020222ca363fac20ea3bc876e68d0b4022e9c75c0539072aa4f7ceec8e3068"} Dec 03 00:47:01 crc kubenswrapper[4912]: I1203 00:47:01.824317 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-bbcd684f4-hp2d8" podStartSLOduration=38.824300743 podStartE2EDuration="38.824300743s" podCreationTimestamp="2025-12-03 00:46:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:01.816629222 +0000 UTC m=+1407.458649802" watchObservedRunningTime="2025-12-03 00:47:01.824300743 +0000 UTC m=+1407.466321303" Dec 03 00:47:02 crc kubenswrapper[4912]: I1203 00:47:02.795787 4912 generic.go:334] "Generic (PLEG): container finished" podID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerID="5f5522f339ab2c5d4dc14dc9af58e23cdb8b198cf2ce18b2e927cc23c4b15657" exitCode=0 Dec 03 00:47:02 crc kubenswrapper[4912]: I1203 00:47:02.795913 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4smn9" event={"ID":"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f","Type":"ContainerDied","Data":"5f5522f339ab2c5d4dc14dc9af58e23cdb8b198cf2ce18b2e927cc23c4b15657"} Dec 03 00:47:04 crc kubenswrapper[4912]: I1203 00:47:04.389731 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:47:04 crc kubenswrapper[4912]: I1203 00:47:04.390152 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:47:04 crc kubenswrapper[4912]: I1203 00:47:04.397102 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:47:04 crc kubenswrapper[4912]: I1203 00:47:04.833087 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e835f70a-04f2-43d3-ba92-6a6d20216a7d","Type":"ContainerStarted","Data":"176f3baafbd5b1f15d895b15b75097eb584a8c8b88714151d5d20d50d88ab55d"} Dec 03 00:47:04 crc kubenswrapper[4912]: I1203 00:47:04.838109 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-bbcd684f4-hp2d8" Dec 03 00:47:04 crc kubenswrapper[4912]: I1203 00:47:04.861065 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371990.99373 podStartE2EDuration="45.861045906s" podCreationTimestamp="2025-12-03 00:46:19 +0000 UTC" firstStartedPulling="2025-12-03 00:46:21.587202863 +0000 UTC m=+1367.229223423" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:04.854546265 +0000 UTC m=+1410.496566835" watchObservedRunningTime="2025-12-03 00:47:04.861045906 +0000 UTC m=+1410.503066456" Dec 03 00:47:04 crc kubenswrapper[4912]: I1203 00:47:04.944533 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5674795f8c-jxb67"] Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.845513 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9a1a55bf-bc0f-4004-b540-ead98e30d25a","Type":"ContainerStarted","Data":"b4733de340110e5cd2f36848058a8c77a4d1919a6825a4329f2d7292fcaf78fe"} Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.850104 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4smn9" event={"ID":"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f","Type":"ContainerStarted","Data":"3f58a122668d0e2ab731a664ea789afbb31c750daa6f21bdb5d0ad54907a9ffa"} Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.852439 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d0dca179-6a1f-41ad-bad5-1d193e2583c8","Type":"ContainerStarted","Data":"00c42f86a3ff162ddbd697eff41a296adf806d4f1d772253fd5d6e0ce48204f5"} Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.855736 4912 generic.go:334] "Generic (PLEG): container finished" podID="2c613cf4-0585-466f-8db2-4c4f4b4765f0" containerID="24ed8406fdf1701ab33b43297da1ce3947e65b0e437fc99f77ec8a775b83a4ef" exitCode=0 Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.855850 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6lm7j" event={"ID":"2c613cf4-0585-466f-8db2-4c4f4b4765f0","Type":"ContainerDied","Data":"24ed8406fdf1701ab33b43297da1ce3947e65b0e437fc99f77ec8a775b83a4ef"} Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.858138 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026","Type":"ContainerStarted","Data":"35b0682d3d517b541e13311751a2dbccc890949ab55feefbc5d6e600a1ecaded"} Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.859737 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" event={"ID":"9d7ffd59-b442-43e2-b7b4-5a2fc42507d3","Type":"ContainerStarted","Data":"16ab4bdbe98631ab46109a7656ad1b91fe59c4203b7adc86959d6cbec958ddc3"} Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.864088 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gbbh7" event={"ID":"dd4efef6-5ce3-444a-a464-1f18f7cc2db4","Type":"ContainerStarted","Data":"7f7c54be3138232774167c2a3bf1f4efe2636fb2e5c0b5885aa193d06e818e50"} Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.864193 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-gbbh7" Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.866721 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d4611c7f-5acd-4fb6-a984-ef56f9feeb72","Type":"ContainerStarted","Data":"081242bfe133d709203bcb84c3ce63df276a39cd8480362677e97e0354cd4a75"} Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.866889 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.879866 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=11.788515269 podStartE2EDuration="48.879843529s" podCreationTimestamp="2025-12-03 00:46:17 +0000 UTC" firstStartedPulling="2025-12-03 00:46:20.151932757 +0000 UTC m=+1365.793953317" lastFinishedPulling="2025-12-03 00:46:57.243261027 +0000 UTC m=+1402.885281577" observedRunningTime="2025-12-03 00:47:05.875214686 +0000 UTC m=+1411.517235256" watchObservedRunningTime="2025-12-03 00:47:05.879843529 +0000 UTC m=+1411.521864079" Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.939076 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=36.537663598 podStartE2EDuration="43.939054376s" podCreationTimestamp="2025-12-03 00:46:22 +0000 UTC" firstStartedPulling="2025-12-03 00:46:57.23233234 +0000 UTC m=+1402.874352900" lastFinishedPulling="2025-12-03 00:47:04.633723118 +0000 UTC m=+1410.275743678" observedRunningTime="2025-12-03 00:47:05.929603507 +0000 UTC m=+1411.571624067" watchObservedRunningTime="2025-12-03 00:47:05.939054376 +0000 UTC m=+1411.581074936" Dec 03 00:47:05 crc kubenswrapper[4912]: I1203 00:47:05.989544 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-gbbh7" podStartSLOduration=32.873001436 podStartE2EDuration="38.989521683s" podCreationTimestamp="2025-12-03 00:46:27 +0000 UTC" firstStartedPulling="2025-12-03 00:46:56.210688492 +0000 UTC m=+1401.852709052" lastFinishedPulling="2025-12-03 00:47:02.327208739 +0000 UTC m=+1407.969229299" observedRunningTime="2025-12-03 00:47:05.989108642 +0000 UTC m=+1411.631129212" watchObservedRunningTime="2025-12-03 00:47:05.989521683 +0000 UTC m=+1411.631542253" Dec 03 00:47:06 crc kubenswrapper[4912]: I1203 00:47:06.007274 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-96rq5" podStartSLOduration=36.332929474 podStartE2EDuration="43.007248159s" podCreationTimestamp="2025-12-03 00:46:23 +0000 UTC" firstStartedPulling="2025-12-03 00:46:57.814687036 +0000 UTC m=+1403.456707586" lastFinishedPulling="2025-12-03 00:47:04.489005711 +0000 UTC m=+1410.131026271" observedRunningTime="2025-12-03 00:47:06.004765414 +0000 UTC m=+1411.646785974" watchObservedRunningTime="2025-12-03 00:47:06.007248159 +0000 UTC m=+1411.649268719" Dec 03 00:47:06 crc kubenswrapper[4912]: I1203 00:47:06.058399 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 00:47:06 crc kubenswrapper[4912]: I1203 00:47:06.881142 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6lm7j" event={"ID":"2c613cf4-0585-466f-8db2-4c4f4b4765f0","Type":"ContainerStarted","Data":"9fa6fbeb5c83f87c1699c2f2b251d85c22b9d04aac74415c88cc4b4a46e59413"} Dec 03 00:47:06 crc kubenswrapper[4912]: I1203 00:47:06.881973 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6lm7j" event={"ID":"2c613cf4-0585-466f-8db2-4c4f4b4765f0","Type":"ContainerStarted","Data":"9554badf88eaad9dbc1e4dffab2bdded5178961dfb2cf6aec5f22e7b9515b369"} Dec 03 00:47:06 crc kubenswrapper[4912]: I1203 00:47:06.882380 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:47:06 crc kubenswrapper[4912]: I1203 00:47:06.882405 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:47:06 crc kubenswrapper[4912]: I1203 00:47:06.889387 4912 generic.go:334] "Generic (PLEG): container finished" podID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerID="3f58a122668d0e2ab731a664ea789afbb31c750daa6f21bdb5d0ad54907a9ffa" exitCode=0 Dec 03 00:47:06 crc kubenswrapper[4912]: I1203 00:47:06.890897 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4smn9" event={"ID":"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f","Type":"ContainerDied","Data":"3f58a122668d0e2ab731a664ea789afbb31c750daa6f21bdb5d0ad54907a9ffa"} Dec 03 00:47:06 crc kubenswrapper[4912]: I1203 00:47:06.927812 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6lm7j" podStartSLOduration=36.422510144 podStartE2EDuration="39.920401514s" podCreationTimestamp="2025-12-03 00:46:27 +0000 UTC" firstStartedPulling="2025-12-03 00:46:58.634247349 +0000 UTC m=+1404.276267909" lastFinishedPulling="2025-12-03 00:47:02.132138719 +0000 UTC m=+1407.774159279" observedRunningTime="2025-12-03 00:47:06.917682292 +0000 UTC m=+1412.559702862" watchObservedRunningTime="2025-12-03 00:47:06.920401514 +0000 UTC m=+1412.562422084" Dec 03 00:47:08 crc kubenswrapper[4912]: I1203 00:47:08.908954 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerStarted","Data":"ad44c075b7f7b30fd0825216a0ab3035fce81f0b1a2ab801fe30d418c4b82483"} Dec 03 00:47:09 crc kubenswrapper[4912]: I1203 00:47:09.463334 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 00:47:09 crc kubenswrapper[4912]: I1203 00:47:09.463843 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 00:47:10 crc kubenswrapper[4912]: I1203 00:47:10.552480 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 00:47:10 crc kubenswrapper[4912]: I1203 00:47:10.552539 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.075502 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-p7c9n"] Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.077133 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.083839 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.094711 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-p7c9n"] Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.179623 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/290c5de7-2853-45ef-9b03-99a0d301d51f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.179989 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/290c5de7-2853-45ef-9b03-99a0d301d51f-config\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.180112 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55cdx\" (UniqueName: \"kubernetes.io/projected/290c5de7-2853-45ef-9b03-99a0d301d51f-kube-api-access-55cdx\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.180227 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/290c5de7-2853-45ef-9b03-99a0d301d51f-ovs-rundir\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.180579 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/290c5de7-2853-45ef-9b03-99a0d301d51f-combined-ca-bundle\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.180728 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/290c5de7-2853-45ef-9b03-99a0d301d51f-ovn-rundir\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.262376 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q4drb"] Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.282820 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/290c5de7-2853-45ef-9b03-99a0d301d51f-ovn-rundir\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.282968 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/290c5de7-2853-45ef-9b03-99a0d301d51f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.283031 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/290c5de7-2853-45ef-9b03-99a0d301d51f-config\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.283078 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55cdx\" (UniqueName: \"kubernetes.io/projected/290c5de7-2853-45ef-9b03-99a0d301d51f-kube-api-access-55cdx\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.283132 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/290c5de7-2853-45ef-9b03-99a0d301d51f-ovs-rundir\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.283191 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/290c5de7-2853-45ef-9b03-99a0d301d51f-combined-ca-bundle\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.283291 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/290c5de7-2853-45ef-9b03-99a0d301d51f-ovn-rundir\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.283702 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/290c5de7-2853-45ef-9b03-99a0d301d51f-ovs-rundir\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.284407 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/290c5de7-2853-45ef-9b03-99a0d301d51f-config\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.288111 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ctnln"] Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.294422 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/290c5de7-2853-45ef-9b03-99a0d301d51f-combined-ca-bundle\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.300366 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/290c5de7-2853-45ef-9b03-99a0d301d51f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.306980 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.313528 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.317692 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ctnln"] Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.337122 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55cdx\" (UniqueName: \"kubernetes.io/projected/290c5de7-2853-45ef-9b03-99a0d301d51f-kube-api-access-55cdx\") pod \"ovn-controller-metrics-p7c9n\" (UID: \"290c5de7-2853-45ef-9b03-99a0d301d51f\") " pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.386629 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj8dm\" (UniqueName: \"kubernetes.io/projected/8280e167-7f99-4855-b886-deb7edfc7cd6-kube-api-access-vj8dm\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.386683 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.386779 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.386824 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-config\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.413159 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-p7c9n" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.492094 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.492179 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-config\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.492250 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj8dm\" (UniqueName: \"kubernetes.io/projected/8280e167-7f99-4855-b886-deb7edfc7cd6-kube-api-access-vj8dm\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.492280 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.493221 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.493431 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-config\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.493555 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.494725 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dt8zv"] Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.532818 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-lphwt"] Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.539403 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.545037 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.545411 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj8dm\" (UniqueName: \"kubernetes.io/projected/8280e167-7f99-4855-b886-deb7edfc7cd6-kube-api-access-vj8dm\") pod \"dnsmasq-dns-5bf47b49b7-ctnln\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.560480 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lphwt"] Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.593569 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-config\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.593618 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l586b\" (UniqueName: \"kubernetes.io/projected/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-kube-api-access-l586b\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.593715 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.593759 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-dns-svc\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.593780 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.696049 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.696164 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-dns-svc\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.696187 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.696264 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-config\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.696979 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l586b\" (UniqueName: \"kubernetes.io/projected/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-kube-api-access-l586b\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.697589 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.698233 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.698563 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-dns-svc\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.699197 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-config\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.720568 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.723080 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l586b\" (UniqueName: \"kubernetes.io/projected/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-kube-api-access-l586b\") pod \"dnsmasq-dns-8554648995-lphwt\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:11 crc kubenswrapper[4912]: I1203 00:47:11.915777 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.744709 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ctnln"] Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.774455 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8rk7b"] Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.776096 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.822872 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.822992 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.823064 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65php\" (UniqueName: \"kubernetes.io/projected/924dacce-93b0-4492-bded-aa0ac2668954-kube-api-access-65php\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.823094 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-config\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.823125 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.827536 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8rk7b"] Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.925133 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.925204 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.925276 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65php\" (UniqueName: \"kubernetes.io/projected/924dacce-93b0-4492-bded-aa0ac2668954-kube-api-access-65php\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.925303 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-config\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.925328 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.926784 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.926855 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.927404 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-config\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.927418 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:12 crc kubenswrapper[4912]: I1203 00:47:12.973444 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65php\" (UniqueName: \"kubernetes.io/projected/924dacce-93b0-4492-bded-aa0ac2668954-kube-api-access-65php\") pod \"dnsmasq-dns-b8fbc5445-8rk7b\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:13 crc kubenswrapper[4912]: I1203 00:47:13.110968 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:13 crc kubenswrapper[4912]: I1203 00:47:13.158857 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 00:47:13 crc kubenswrapper[4912]: I1203 00:47:13.917274 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 00:47:13 crc kubenswrapper[4912]: I1203 00:47:13.924303 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 00:47:13 crc kubenswrapper[4912]: I1203 00:47:13.928833 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 00:47:13 crc kubenswrapper[4912]: I1203 00:47:13.928890 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 00:47:13 crc kubenswrapper[4912]: I1203 00:47:13.928948 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-xfq4h" Dec 03 00:47:13 crc kubenswrapper[4912]: I1203 00:47:13.929014 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 00:47:13 crc kubenswrapper[4912]: I1203 00:47:13.954902 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.049193 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz5sn\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-kube-api-access-tz5sn\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.049643 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-cache\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.049942 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.050109 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-lock\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.050227 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.152652 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.152730 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-lock\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.152761 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.152797 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz5sn\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-kube-api-access-tz5sn\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.152832 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-cache\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: E1203 00:47:14.152928 4912 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 00:47:14 crc kubenswrapper[4912]: E1203 00:47:14.152962 4912 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 00:47:14 crc kubenswrapper[4912]: E1203 00:47:14.153044 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift podName:761ebfbd-144d-42cc-ac4f-09a0c49ec7bb nodeName:}" failed. No retries permitted until 2025-12-03 00:47:14.653016803 +0000 UTC m=+1420.295037363 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift") pod "swift-storage-0" (UID: "761ebfbd-144d-42cc-ac4f-09a0c49ec7bb") : configmap "swift-ring-files" not found Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.153337 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.153643 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-cache\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.162380 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-lock\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.177659 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz5sn\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-kube-api-access-tz5sn\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.192609 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.663070 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:14 crc kubenswrapper[4912]: E1203 00:47:14.663362 4912 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 00:47:14 crc kubenswrapper[4912]: E1203 00:47:14.663390 4912 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 00:47:14 crc kubenswrapper[4912]: E1203 00:47:14.663470 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift podName:761ebfbd-144d-42cc-ac4f-09a0c49ec7bb nodeName:}" failed. No retries permitted until 2025-12-03 00:47:15.663433176 +0000 UTC m=+1421.305453736 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift") pod "swift-storage-0" (UID: "761ebfbd-144d-42cc-ac4f-09a0c49ec7bb") : configmap "swift-ring-files" not found Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.664953 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.765837 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-config\") pod \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.766443 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t654h\" (UniqueName: \"kubernetes.io/projected/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-kube-api-access-t654h\") pod \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.766489 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-dns-svc\") pod \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\" (UID: \"4cd22f06-6fc2-4edd-b0fc-44f4307878f0\") " Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.766790 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-config" (OuterVolumeSpecName: "config") pod "4cd22f06-6fc2-4edd-b0fc-44f4307878f0" (UID: "4cd22f06-6fc2-4edd-b0fc-44f4307878f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.767601 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4cd22f06-6fc2-4edd-b0fc-44f4307878f0" (UID: "4cd22f06-6fc2-4edd-b0fc-44f4307878f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.770293 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.770318 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.771242 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-kube-api-access-t654h" (OuterVolumeSpecName: "kube-api-access-t654h") pod "4cd22f06-6fc2-4edd-b0fc-44f4307878f0" (UID: "4cd22f06-6fc2-4edd-b0fc-44f4307878f0"). InnerVolumeSpecName "kube-api-access-t654h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.871671 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t654h\" (UniqueName: \"kubernetes.io/projected/4cd22f06-6fc2-4edd-b0fc-44f4307878f0-kube-api-access-t654h\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.972916 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" event={"ID":"4cd22f06-6fc2-4edd-b0fc-44f4307878f0","Type":"ContainerDied","Data":"eeb0ab237e8d2a20bc9f8e7166aa8afae8c54208108ffc3467f5f1d52545c43d"} Dec 03 00:47:14 crc kubenswrapper[4912]: I1203 00:47:14.973029 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-dt8zv" Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.001896 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4smn9" event={"ID":"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f","Type":"ContainerStarted","Data":"bcfc7bca6c4259cc09d7cbb6628dba16e94f746d8e12e4e611ccfd15d71d4f36"} Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.027425 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4smn9" podStartSLOduration=34.996535621 podStartE2EDuration="40.027395417s" podCreationTimestamp="2025-12-03 00:46:35 +0000 UTC" firstStartedPulling="2025-12-03 00:47:02.930699 +0000 UTC m=+1408.572719560" lastFinishedPulling="2025-12-03 00:47:07.961558796 +0000 UTC m=+1413.603579356" observedRunningTime="2025-12-03 00:47:15.026563775 +0000 UTC m=+1420.668584345" watchObservedRunningTime="2025-12-03 00:47:15.027395417 +0000 UTC m=+1420.669415977" Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.125256 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dt8zv"] Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.142282 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-dt8zv"] Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.252524 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lphwt"] Dec 03 00:47:15 crc kubenswrapper[4912]: W1203 00:47:15.254829 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d0c1d3b_5d5d_4d26_bc5b_3a6dddd5acba.slice/crio-a247c66dac65b4ee2675b27955080f648d953fd1c63fa7a6c00aab0aa083ee9a WatchSource:0}: Error finding container a247c66dac65b4ee2675b27955080f648d953fd1c63fa7a6c00aab0aa083ee9a: Status 404 returned error can't find the container with id a247c66dac65b4ee2675b27955080f648d953fd1c63fa7a6c00aab0aa083ee9a Dec 03 00:47:15 crc kubenswrapper[4912]: W1203 00:47:15.389261 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod924dacce_93b0_4492_bded_aa0ac2668954.slice/crio-d3e93ff36465134334c09c375a47cdb574dc1b3b52caa4c6977ece9016b6c1b8 WatchSource:0}: Error finding container d3e93ff36465134334c09c375a47cdb574dc1b3b52caa4c6977ece9016b6c1b8: Status 404 returned error can't find the container with id d3e93ff36465134334c09c375a47cdb574dc1b3b52caa4c6977ece9016b6c1b8 Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.437481 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8rk7b"] Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.465399 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-p7c9n"] Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.542706 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ctnln"] Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.731658 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:15 crc kubenswrapper[4912]: E1203 00:47:15.731918 4912 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 00:47:15 crc kubenswrapper[4912]: E1203 00:47:15.731958 4912 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 00:47:15 crc kubenswrapper[4912]: E1203 00:47:15.732047 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift podName:761ebfbd-144d-42cc-ac4f-09a0c49ec7bb nodeName:}" failed. No retries permitted until 2025-12-03 00:47:17.732019818 +0000 UTC m=+1423.374040378 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift") pod "swift-storage-0" (UID: "761ebfbd-144d-42cc-ac4f-09a0c49ec7bb") : configmap "swift-ring-files" not found Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.781870 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:47:15 crc kubenswrapper[4912]: I1203 00:47:15.781952 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.043894 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026","Type":"ContainerStarted","Data":"680d682c3f36ead9a3e764ab6cfe6c6a8d00df10bf423245ec819a13117dc7e9"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.051309 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d0dca179-6a1f-41ad-bad5-1d193e2583c8","Type":"ContainerStarted","Data":"6b9175fe1f31cdeb184801f2f9e46ee7dd0453ef0db5102cbe00dce7b4e83d22"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.063303 4912 generic.go:334] "Generic (PLEG): container finished" podID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerID="ad44c075b7f7b30fd0825216a0ab3035fce81f0b1a2ab801fe30d418c4b82483" exitCode=0 Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.063585 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerDied","Data":"ad44c075b7f7b30fd0825216a0ab3035fce81f0b1a2ab801fe30d418c4b82483"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.065408 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-p7c9n" event={"ID":"290c5de7-2853-45ef-9b03-99a0d301d51f","Type":"ContainerStarted","Data":"af319cca3112bfc685c5bc19a8ae0045c2388fc2e439282bba988c77fd1cf019"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.081242 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=35.752058065 podStartE2EDuration="51.081219392s" podCreationTimestamp="2025-12-03 00:46:25 +0000 UTC" firstStartedPulling="2025-12-03 00:46:59.379509268 +0000 UTC m=+1405.021529838" lastFinishedPulling="2025-12-03 00:47:14.708670605 +0000 UTC m=+1420.350691165" observedRunningTime="2025-12-03 00:47:16.079078965 +0000 UTC m=+1421.721099525" watchObservedRunningTime="2025-12-03 00:47:16.081219392 +0000 UTC m=+1421.723239952" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.090299 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" event={"ID":"924dacce-93b0-4492-bded-aa0ac2668954","Type":"ContainerStarted","Data":"85331a75effc382414de473f18aa352fb989a2da8d27a27c1f8d5b8c4c82e321"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.090464 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" event={"ID":"924dacce-93b0-4492-bded-aa0ac2668954","Type":"ContainerStarted","Data":"d3e93ff36465134334c09c375a47cdb574dc1b3b52caa4c6977ece9016b6c1b8"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.108996 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" event={"ID":"8280e167-7f99-4855-b886-deb7edfc7cd6","Type":"ContainerStarted","Data":"1945e58612a313556fdc712f617a6ea6b8a596e63f73da3af3e7eb97fbca823f"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.109302 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" podUID="8280e167-7f99-4855-b886-deb7edfc7cd6" containerName="init" containerID="cri-o://147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306" gracePeriod=10 Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.111998 4912 generic.go:334] "Generic (PLEG): container finished" podID="40178e9e-c6a8-4742-b5db-f090c3f462e1" containerID="6b69e1cf94ee6136166a5c104036a8e21d211c9e6f4d62ddbb459e6fd9bf341a" exitCode=0 Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.112045 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-q4drb" event={"ID":"40178e9e-c6a8-4742-b5db-f090c3f462e1","Type":"ContainerDied","Data":"6b69e1cf94ee6136166a5c104036a8e21d211c9e6f4d62ddbb459e6fd9bf341a"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.131399 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lphwt" event={"ID":"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba","Type":"ContainerStarted","Data":"e09847018f67cff28ea0a2e1b264c159721c31ed141fc0a7d26cc92f37cd7639"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.131476 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lphwt" event={"ID":"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba","Type":"ContainerStarted","Data":"a247c66dac65b4ee2675b27955080f648d953fd1c63fa7a6c00aab0aa083ee9a"} Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.186939 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=31.006546608 podStartE2EDuration="47.186919281s" podCreationTimestamp="2025-12-03 00:46:29 +0000 UTC" firstStartedPulling="2025-12-03 00:46:58.60387447 +0000 UTC m=+1404.245895030" lastFinishedPulling="2025-12-03 00:47:14.784247143 +0000 UTC m=+1420.426267703" observedRunningTime="2025-12-03 00:47:16.185059413 +0000 UTC m=+1421.827079973" watchObservedRunningTime="2025-12-03 00:47:16.186919281 +0000 UTC m=+1421.828939841" Dec 03 00:47:16 crc kubenswrapper[4912]: E1203 00:47:16.579751 4912 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 03 00:47:16 crc kubenswrapper[4912]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 00:47:16 crc kubenswrapper[4912]: > podSandboxID="a247c66dac65b4ee2675b27955080f648d953fd1c63fa7a6c00aab0aa083ee9a" Dec 03 00:47:16 crc kubenswrapper[4912]: E1203 00:47:16.580271 4912 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 03 00:47:16 crc kubenswrapper[4912]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n654h99h64ch5dbh6dh555h587h64bh5cfh647h5fdh57ch679h9h597h5f5hbch59bh54fh575h566h667h586h5f5h65ch5bch57h68h65ch58bh694h5cfq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l586b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-8554648995-lphwt_openstack(0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 00:47:16 crc kubenswrapper[4912]: > logger="UnhandledError" Dec 03 00:47:16 crc kubenswrapper[4912]: E1203 00:47:16.581518 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-8554648995-lphwt" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.600094 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cd22f06-6fc2-4edd-b0fc-44f4307878f0" path="/var/lib/kubelet/pods/4cd22f06-6fc2-4edd-b0fc-44f4307878f0/volumes" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.649991 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.759321 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6n7p\" (UniqueName: \"kubernetes.io/projected/40178e9e-c6a8-4742-b5db-f090c3f462e1-kube-api-access-l6n7p\") pod \"40178e9e-c6a8-4742-b5db-f090c3f462e1\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.759491 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-config\") pod \"40178e9e-c6a8-4742-b5db-f090c3f462e1\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.759529 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-dns-svc\") pod \"40178e9e-c6a8-4742-b5db-f090c3f462e1\" (UID: \"40178e9e-c6a8-4742-b5db-f090c3f462e1\") " Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.769326 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40178e9e-c6a8-4742-b5db-f090c3f462e1-kube-api-access-l6n7p" (OuterVolumeSpecName: "kube-api-access-l6n7p") pod "40178e9e-c6a8-4742-b5db-f090c3f462e1" (UID: "40178e9e-c6a8-4742-b5db-f090c3f462e1"). InnerVolumeSpecName "kube-api-access-l6n7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.771691 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.789075 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-config" (OuterVolumeSpecName: "config") pod "40178e9e-c6a8-4742-b5db-f090c3f462e1" (UID: "40178e9e-c6a8-4742-b5db-f090c3f462e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.803101 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "40178e9e-c6a8-4742-b5db-f090c3f462e1" (UID: "40178e9e-c6a8-4742-b5db-f090c3f462e1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.861377 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-config\") pod \"8280e167-7f99-4855-b886-deb7edfc7cd6\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.861525 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-ovsdbserver-nb\") pod \"8280e167-7f99-4855-b886-deb7edfc7cd6\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.861750 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-dns-svc\") pod \"8280e167-7f99-4855-b886-deb7edfc7cd6\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.861804 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj8dm\" (UniqueName: \"kubernetes.io/projected/8280e167-7f99-4855-b886-deb7edfc7cd6-kube-api-access-vj8dm\") pod \"8280e167-7f99-4855-b886-deb7edfc7cd6\" (UID: \"8280e167-7f99-4855-b886-deb7edfc7cd6\") " Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.862369 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6n7p\" (UniqueName: \"kubernetes.io/projected/40178e9e-c6a8-4742-b5db-f090c3f462e1-kube-api-access-l6n7p\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.862386 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.862395 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/40178e9e-c6a8-4742-b5db-f090c3f462e1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.873750 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8280e167-7f99-4855-b886-deb7edfc7cd6-kube-api-access-vj8dm" (OuterVolumeSpecName: "kube-api-access-vj8dm") pod "8280e167-7f99-4855-b886-deb7edfc7cd6" (UID: "8280e167-7f99-4855-b886-deb7edfc7cd6"). InnerVolumeSpecName "kube-api-access-vj8dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.895730 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8280e167-7f99-4855-b886-deb7edfc7cd6" (UID: "8280e167-7f99-4855-b886-deb7edfc7cd6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.901866 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-config" (OuterVolumeSpecName: "config") pod "8280e167-7f99-4855-b886-deb7edfc7cd6" (UID: "8280e167-7f99-4855-b886-deb7edfc7cd6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.902755 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8280e167-7f99-4855-b886-deb7edfc7cd6" (UID: "8280e167-7f99-4855-b886-deb7edfc7cd6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.916878 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.949237 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4smn9" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerName="registry-server" probeResult="failure" output=< Dec 03 00:47:16 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 00:47:16 crc kubenswrapper[4912]: > Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.964297 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.964346 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj8dm\" (UniqueName: \"kubernetes.io/projected/8280e167-7f99-4855-b886-deb7edfc7cd6-kube-api-access-vj8dm\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.964356 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:16 crc kubenswrapper[4912]: I1203 00:47:16.964364 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8280e167-7f99-4855-b886-deb7edfc7cd6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.065930 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.143286 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-q4drb" event={"ID":"40178e9e-c6a8-4742-b5db-f090c3f462e1","Type":"ContainerDied","Data":"e071d984f52c1a9333b0786f0ba7e20f7bef72a070f905815fa954b966d398c2"} Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.143357 4912 scope.go:117] "RemoveContainer" containerID="6b69e1cf94ee6136166a5c104036a8e21d211c9e6f4d62ddbb459e6fd9bf341a" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.143864 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-q4drb" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.155584 4912 generic.go:334] "Generic (PLEG): container finished" podID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerID="e09847018f67cff28ea0a2e1b264c159721c31ed141fc0a7d26cc92f37cd7639" exitCode=0 Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.155661 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lphwt" event={"ID":"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba","Type":"ContainerDied","Data":"e09847018f67cff28ea0a2e1b264c159721c31ed141fc0a7d26cc92f37cd7639"} Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.158518 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-p7c9n" event={"ID":"290c5de7-2853-45ef-9b03-99a0d301d51f","Type":"ContainerStarted","Data":"a8219cabc7dc514c527aa6af987f5405bee999d17cb18ff50286c814478824a5"} Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.170066 4912 generic.go:334] "Generic (PLEG): container finished" podID="924dacce-93b0-4492-bded-aa0ac2668954" containerID="85331a75effc382414de473f18aa352fb989a2da8d27a27c1f8d5b8c4c82e321" exitCode=0 Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.170504 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" event={"ID":"924dacce-93b0-4492-bded-aa0ac2668954","Type":"ContainerDied","Data":"85331a75effc382414de473f18aa352fb989a2da8d27a27c1f8d5b8c4c82e321"} Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.170586 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" event={"ID":"924dacce-93b0-4492-bded-aa0ac2668954","Type":"ContainerStarted","Data":"ff42eb77a8a089f3a34825f7746e742d10e005083c6e47ecb9b181a64c56c546"} Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.171033 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.172780 4912 generic.go:334] "Generic (PLEG): container finished" podID="8280e167-7f99-4855-b886-deb7edfc7cd6" containerID="147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306" exitCode=0 Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.173197 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" event={"ID":"8280e167-7f99-4855-b886-deb7edfc7cd6","Type":"ContainerDied","Data":"1945e58612a313556fdc712f617a6ea6b8a596e63f73da3af3e7eb97fbca823f"} Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.173252 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" event={"ID":"8280e167-7f99-4855-b886-deb7edfc7cd6","Type":"ContainerDied","Data":"147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306"} Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.173321 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-ctnln" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.216872 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-p7c9n" podStartSLOduration=6.216845137 podStartE2EDuration="6.216845137s" podCreationTimestamp="2025-12-03 00:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:17.213244652 +0000 UTC m=+1422.855265222" watchObservedRunningTime="2025-12-03 00:47:17.216845137 +0000 UTC m=+1422.858865697" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.228150 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.267281 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" podStartSLOduration=5.267255623 podStartE2EDuration="5.267255623s" podCreationTimestamp="2025-12-03 00:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:17.251112318 +0000 UTC m=+1422.893132878" watchObservedRunningTime="2025-12-03 00:47:17.267255623 +0000 UTC m=+1422.909276183" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.303663 4912 scope.go:117] "RemoveContainer" containerID="147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.379528 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q4drb"] Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.431670 4912 scope.go:117] "RemoveContainer" containerID="147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306" Dec 03 00:47:17 crc kubenswrapper[4912]: E1203 00:47:17.440671 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306\": container with ID starting with 147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306 not found: ID does not exist" containerID="147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.440747 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306"} err="failed to get container status \"147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306\": rpc error: code = NotFound desc = could not find container \"147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306\": container with ID starting with 147effde5eedcfb64354e8d428ea4f9d88e9751c1e0ff8f042bebdb90ab40306 not found: ID does not exist" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.442621 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-q4drb"] Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.468074 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ctnln"] Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.484655 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ctnln"] Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.807540 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:17 crc kubenswrapper[4912]: E1203 00:47:17.807775 4912 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 00:47:17 crc kubenswrapper[4912]: E1203 00:47:17.807831 4912 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 00:47:17 crc kubenswrapper[4912]: E1203 00:47:17.807923 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift podName:761ebfbd-144d-42cc-ac4f-09a0c49ec7bb nodeName:}" failed. No retries permitted until 2025-12-03 00:47:21.807896211 +0000 UTC m=+1427.449916761 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift") pod "swift-storage-0" (UID: "761ebfbd-144d-42cc-ac4f-09a0c49ec7bb") : configmap "swift-ring-files" not found Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.859938 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6wv8k"] Dec 03 00:47:17 crc kubenswrapper[4912]: E1203 00:47:17.860400 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40178e9e-c6a8-4742-b5db-f090c3f462e1" containerName="init" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.860449 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="40178e9e-c6a8-4742-b5db-f090c3f462e1" containerName="init" Dec 03 00:47:17 crc kubenswrapper[4912]: E1203 00:47:17.860493 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8280e167-7f99-4855-b886-deb7edfc7cd6" containerName="init" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.860500 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8280e167-7f99-4855-b886-deb7edfc7cd6" containerName="init" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.860693 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="40178e9e-c6a8-4742-b5db-f090c3f462e1" containerName="init" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.860714 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="8280e167-7f99-4855-b886-deb7edfc7cd6" containerName="init" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.866552 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.909799 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-scripts\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.909906 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1556422b-130c-4bf1-8010-ee7c6ea5f662-etc-swift\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.909947 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-ring-data-devices\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.909973 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-dispersionconf\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.910021 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-swiftconf\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.910043 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.910059 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkksb\" (UniqueName: \"kubernetes.io/projected/1556422b-130c-4bf1-8010-ee7c6ea5f662-kube-api-access-zkksb\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.910237 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.910613 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-combined-ca-bundle\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.910348 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 00:47:17 crc kubenswrapper[4912]: I1203 00:47:17.912775 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6wv8k"] Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.013626 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-ring-data-devices\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.013739 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-dispersionconf\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.013827 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-swiftconf\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.013900 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkksb\" (UniqueName: \"kubernetes.io/projected/1556422b-130c-4bf1-8010-ee7c6ea5f662-kube-api-access-zkksb\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.014079 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-combined-ca-bundle\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.014206 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-scripts\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.014302 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1556422b-130c-4bf1-8010-ee7c6ea5f662-etc-swift\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.015110 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1556422b-130c-4bf1-8010-ee7c6ea5f662-etc-swift\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.015921 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-ring-data-devices\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.019187 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-scripts\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.022658 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-combined-ca-bundle\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.034875 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-dispersionconf\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.036937 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-swiftconf\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.058234 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkksb\" (UniqueName: \"kubernetes.io/projected/1556422b-130c-4bf1-8010-ee7c6ea5f662-kube-api-access-zkksb\") pod \"swift-ring-rebalance-6wv8k\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.189058 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.213753 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lphwt" event={"ID":"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba","Type":"ContainerStarted","Data":"dd17c1a35e8eb97fc1b40aee9145526c2e65db13c38cd7105d30888849a8ab1b"} Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.213832 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.231739 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.263898 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-lphwt" podStartSLOduration=7.2638730129999995 podStartE2EDuration="7.263873013s" podCreationTimestamp="2025-12-03 00:47:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:18.243940598 +0000 UTC m=+1423.885961168" watchObservedRunningTime="2025-12-03 00:47:18.263873013 +0000 UTC m=+1423.905893573" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.293129 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.585985 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40178e9e-c6a8-4742-b5db-f090c3f462e1" path="/var/lib/kubelet/pods/40178e9e-c6a8-4742-b5db-f090c3f462e1/volumes" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.586961 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8280e167-7f99-4855-b886-deb7edfc7cd6" path="/var/lib/kubelet/pods/8280e167-7f99-4855-b886-deb7edfc7cd6/volumes" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.641108 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.709427 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 00:47:18 crc kubenswrapper[4912]: I1203 00:47:18.727462 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6wv8k"] Dec 03 00:47:18 crc kubenswrapper[4912]: W1203 00:47:18.746401 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1556422b_130c_4bf1_8010_ee7c6ea5f662.slice/crio-74490d463f719b89303f16e0bf99043fe1efd7d28d1b6c3e46248505c6a1358f WatchSource:0}: Error finding container 74490d463f719b89303f16e0bf99043fe1efd7d28d1b6c3e46248505c6a1358f: Status 404 returned error can't find the container with id 74490d463f719b89303f16e0bf99043fe1efd7d28d1b6c3e46248505c6a1358f Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.234772 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6wv8k" event={"ID":"1556422b-130c-4bf1-8010-ee7c6ea5f662","Type":"ContainerStarted","Data":"74490d463f719b89303f16e0bf99043fe1efd7d28d1b6c3e46248505c6a1358f"} Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.235472 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.289618 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.299002 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.414150 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.542086 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.577975 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.579845 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.582550 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.582767 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.594359 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-pwklf" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.599649 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.609568 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.687137 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/197a0bff-4ebc-4581-bc0f-77d459500339-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.687224 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/197a0bff-4ebc-4581-bc0f-77d459500339-config\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.687337 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/197a0bff-4ebc-4581-bc0f-77d459500339-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.687414 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197a0bff-4ebc-4581-bc0f-77d459500339-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.687524 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/197a0bff-4ebc-4581-bc0f-77d459500339-scripts\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.687587 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k64c8\" (UniqueName: \"kubernetes.io/projected/197a0bff-4ebc-4581-bc0f-77d459500339-kube-api-access-k64c8\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.687607 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/197a0bff-4ebc-4581-bc0f-77d459500339-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.789719 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k64c8\" (UniqueName: \"kubernetes.io/projected/197a0bff-4ebc-4581-bc0f-77d459500339-kube-api-access-k64c8\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.789775 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/197a0bff-4ebc-4581-bc0f-77d459500339-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.789838 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/197a0bff-4ebc-4581-bc0f-77d459500339-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.789897 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/197a0bff-4ebc-4581-bc0f-77d459500339-config\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.789948 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/197a0bff-4ebc-4581-bc0f-77d459500339-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.790013 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197a0bff-4ebc-4581-bc0f-77d459500339-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.790048 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/197a0bff-4ebc-4581-bc0f-77d459500339-scripts\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.790979 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/197a0bff-4ebc-4581-bc0f-77d459500339-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.791284 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/197a0bff-4ebc-4581-bc0f-77d459500339-scripts\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.792789 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/197a0bff-4ebc-4581-bc0f-77d459500339-config\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.798422 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/197a0bff-4ebc-4581-bc0f-77d459500339-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.807250 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/197a0bff-4ebc-4581-bc0f-77d459500339-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.820188 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/197a0bff-4ebc-4581-bc0f-77d459500339-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.826131 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k64c8\" (UniqueName: \"kubernetes.io/projected/197a0bff-4ebc-4581-bc0f-77d459500339-kube-api-access-k64c8\") pod \"ovn-northd-0\" (UID: \"197a0bff-4ebc-4581-bc0f-77d459500339\") " pod="openstack/ovn-northd-0" Dec 03 00:47:19 crc kubenswrapper[4912]: I1203 00:47:19.929747 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.441917 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-8dng2"] Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.443851 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.463946 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8dng2"] Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.490045 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-1612-account-create-update-qg8wc"] Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.491908 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.498735 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.518424 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-operator-scripts\") pod \"keystone-1612-account-create-update-qg8wc\" (UID: \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\") " pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.518593 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e4aa219-c983-4ed9-96e6-e6583d25d34d-operator-scripts\") pod \"keystone-db-create-8dng2\" (UID: \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\") " pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.518644 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp77p\" (UniqueName: \"kubernetes.io/projected/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-kube-api-access-vp77p\") pod \"keystone-1612-account-create-update-qg8wc\" (UID: \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\") " pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.518699 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9qg2\" (UniqueName: \"kubernetes.io/projected/8e4aa219-c983-4ed9-96e6-e6583d25d34d-kube-api-access-d9qg2\") pod \"keystone-db-create-8dng2\" (UID: \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\") " pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.611990 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1612-account-create-update-qg8wc"] Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.622210 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e4aa219-c983-4ed9-96e6-e6583d25d34d-operator-scripts\") pod \"keystone-db-create-8dng2\" (UID: \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\") " pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.622290 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp77p\" (UniqueName: \"kubernetes.io/projected/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-kube-api-access-vp77p\") pod \"keystone-1612-account-create-update-qg8wc\" (UID: \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\") " pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.622359 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9qg2\" (UniqueName: \"kubernetes.io/projected/8e4aa219-c983-4ed9-96e6-e6583d25d34d-kube-api-access-d9qg2\") pod \"keystone-db-create-8dng2\" (UID: \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\") " pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.622510 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-operator-scripts\") pod \"keystone-1612-account-create-update-qg8wc\" (UID: \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\") " pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.624251 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e4aa219-c983-4ed9-96e6-e6583d25d34d-operator-scripts\") pod \"keystone-db-create-8dng2\" (UID: \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\") " pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.624781 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-operator-scripts\") pod \"keystone-1612-account-create-update-qg8wc\" (UID: \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\") " pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.666302 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.686451 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp77p\" (UniqueName: \"kubernetes.io/projected/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-kube-api-access-vp77p\") pod \"keystone-1612-account-create-update-qg8wc\" (UID: \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\") " pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.687681 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9qg2\" (UniqueName: \"kubernetes.io/projected/8e4aa219-c983-4ed9-96e6-e6583d25d34d-kube-api-access-d9qg2\") pod \"keystone-db-create-8dng2\" (UID: \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\") " pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.703832 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-s88vh"] Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.705236 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-s88vh" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.726416 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfv5z\" (UniqueName: \"kubernetes.io/projected/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-kube-api-access-jfv5z\") pod \"placement-db-create-s88vh\" (UID: \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\") " pod="openstack/placement-db-create-s88vh" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.726485 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-operator-scripts\") pod \"placement-db-create-s88vh\" (UID: \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\") " pod="openstack/placement-db-create-s88vh" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.731743 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-s88vh"] Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.796738 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-51b0-account-create-update-vqt9z"] Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.798138 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.804290 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.811407 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-51b0-account-create-update-vqt9z"] Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.829823 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.838869 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53e2fd-3f5c-4382-a815-5e92e3299249-operator-scripts\") pod \"placement-51b0-account-create-update-vqt9z\" (UID: \"ec53e2fd-3f5c-4382-a815-5e92e3299249\") " pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.839016 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpksk\" (UniqueName: \"kubernetes.io/projected/ec53e2fd-3f5c-4382-a815-5e92e3299249-kube-api-access-xpksk\") pod \"placement-51b0-account-create-update-vqt9z\" (UID: \"ec53e2fd-3f5c-4382-a815-5e92e3299249\") " pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.839064 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfv5z\" (UniqueName: \"kubernetes.io/projected/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-kube-api-access-jfv5z\") pod \"placement-db-create-s88vh\" (UID: \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\") " pod="openstack/placement-db-create-s88vh" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.839096 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-operator-scripts\") pod \"placement-db-create-s88vh\" (UID: \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\") " pod="openstack/placement-db-create-s88vh" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.840112 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-operator-scripts\") pod \"placement-db-create-s88vh\" (UID: \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\") " pod="openstack/placement-db-create-s88vh" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.870638 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfv5z\" (UniqueName: \"kubernetes.io/projected/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-kube-api-access-jfv5z\") pod \"placement-db-create-s88vh\" (UID: \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\") " pod="openstack/placement-db-create-s88vh" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.896973 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.940740 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53e2fd-3f5c-4382-a815-5e92e3299249-operator-scripts\") pod \"placement-51b0-account-create-update-vqt9z\" (UID: \"ec53e2fd-3f5c-4382-a815-5e92e3299249\") " pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.940841 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpksk\" (UniqueName: \"kubernetes.io/projected/ec53e2fd-3f5c-4382-a815-5e92e3299249-kube-api-access-xpksk\") pod \"placement-51b0-account-create-update-vqt9z\" (UID: \"ec53e2fd-3f5c-4382-a815-5e92e3299249\") " pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.942835 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53e2fd-3f5c-4382-a815-5e92e3299249-operator-scripts\") pod \"placement-51b0-account-create-update-vqt9z\" (UID: \"ec53e2fd-3f5c-4382-a815-5e92e3299249\") " pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:20 crc kubenswrapper[4912]: I1203 00:47:20.962701 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpksk\" (UniqueName: \"kubernetes.io/projected/ec53e2fd-3f5c-4382-a815-5e92e3299249-kube-api-access-xpksk\") pod \"placement-51b0-account-create-update-vqt9z\" (UID: \"ec53e2fd-3f5c-4382-a815-5e92e3299249\") " pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:21 crc kubenswrapper[4912]: I1203 00:47:21.098511 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-s88vh" Dec 03 00:47:21 crc kubenswrapper[4912]: I1203 00:47:21.264819 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:21 crc kubenswrapper[4912]: I1203 00:47:21.302228 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"197a0bff-4ebc-4581-bc0f-77d459500339","Type":"ContainerStarted","Data":"3b4472365e6ef87c7f98ea5a19b918bb41b441565289c7ebe2c78c33ccf8cb49"} Dec 03 00:47:21 crc kubenswrapper[4912]: I1203 00:47:21.593014 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-8dng2"] Dec 03 00:47:21 crc kubenswrapper[4912]: I1203 00:47:21.873087 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:21 crc kubenswrapper[4912]: E1203 00:47:21.873417 4912 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 00:47:21 crc kubenswrapper[4912]: E1203 00:47:21.873456 4912 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 00:47:21 crc kubenswrapper[4912]: E1203 00:47:21.873532 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift podName:761ebfbd-144d-42cc-ac4f-09a0c49ec7bb nodeName:}" failed. No retries permitted until 2025-12-03 00:47:29.873512717 +0000 UTC m=+1435.515533277 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift") pod "swift-storage-0" (UID: "761ebfbd-144d-42cc-ac4f-09a0c49ec7bb") : configmap "swift-ring-files" not found Dec 03 00:47:21 crc kubenswrapper[4912]: I1203 00:47:21.930150 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-s88vh"] Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.003900 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-1612-account-create-update-qg8wc"] Dec 03 00:47:22 crc kubenswrapper[4912]: W1203 00:47:22.013222 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a43da64_c71a_4ae0_af3a_bc0fb42538ee.slice/crio-9d4d42bb18bbe64ed602b9de7e877100b774603d9a50be259d494cca77ddcb02 WatchSource:0}: Error finding container 9d4d42bb18bbe64ed602b9de7e877100b774603d9a50be259d494cca77ddcb02: Status 404 returned error can't find the container with id 9d4d42bb18bbe64ed602b9de7e877100b774603d9a50be259d494cca77ddcb02 Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.331217 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-51b0-account-create-update-vqt9z"] Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.342655 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1612-account-create-update-qg8wc" event={"ID":"5a43da64-c71a-4ae0-af3a-bc0fb42538ee","Type":"ContainerStarted","Data":"9d4d42bb18bbe64ed602b9de7e877100b774603d9a50be259d494cca77ddcb02"} Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.344957 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8dng2" event={"ID":"8e4aa219-c983-4ed9-96e6-e6583d25d34d","Type":"ContainerStarted","Data":"ed96cd4a41334ddcb073b560b56f7f0e177e7778a628d519295e9798d5dbe3b3"} Dec 03 00:47:22 crc kubenswrapper[4912]: W1203 00:47:22.345115 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec53e2fd_3f5c_4382_a815_5e92e3299249.slice/crio-88905b0fe0bc158f36b00282243c9403cdf74ff0697e10a04c2091b5bf8c38e2 WatchSource:0}: Error finding container 88905b0fe0bc158f36b00282243c9403cdf74ff0697e10a04c2091b5bf8c38e2: Status 404 returned error can't find the container with id 88905b0fe0bc158f36b00282243c9403cdf74ff0697e10a04c2091b5bf8c38e2 Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.346792 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-s88vh" event={"ID":"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd","Type":"ContainerStarted","Data":"c67d8792e77507b7665e93ce63c51cd71fc7662e6a36f5c31a65d95aa4b31b29"} Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.713941 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-htd7z"] Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.716229 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.748793 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-htd7z"] Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.816681 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jm2m\" (UniqueName: \"kubernetes.io/projected/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-kube-api-access-4jm2m\") pod \"mysqld-exporter-openstack-db-create-htd7z\" (UID: \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\") " pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.816755 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-htd7z\" (UID: \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\") " pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.835036 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-91cb-account-create-update-7fmk8"] Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.837075 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.841338 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.854128 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-91cb-account-create-update-7fmk8"] Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.921024 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jm2m\" (UniqueName: \"kubernetes.io/projected/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-kube-api-access-4jm2m\") pod \"mysqld-exporter-openstack-db-create-htd7z\" (UID: \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\") " pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.921163 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-htd7z\" (UID: \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\") " pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.921206 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvmr9\" (UniqueName: \"kubernetes.io/projected/480a1412-80ec-4d41-bd91-ce994043438f-kube-api-access-zvmr9\") pod \"mysqld-exporter-91cb-account-create-update-7fmk8\" (UID: \"480a1412-80ec-4d41-bd91-ce994043438f\") " pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.921293 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/480a1412-80ec-4d41-bd91-ce994043438f-operator-scripts\") pod \"mysqld-exporter-91cb-account-create-update-7fmk8\" (UID: \"480a1412-80ec-4d41-bd91-ce994043438f\") " pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.922623 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-htd7z\" (UID: \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\") " pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:22 crc kubenswrapper[4912]: I1203 00:47:22.947767 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jm2m\" (UniqueName: \"kubernetes.io/projected/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-kube-api-access-4jm2m\") pod \"mysqld-exporter-openstack-db-create-htd7z\" (UID: \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\") " pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.028602 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/480a1412-80ec-4d41-bd91-ce994043438f-operator-scripts\") pod \"mysqld-exporter-91cb-account-create-update-7fmk8\" (UID: \"480a1412-80ec-4d41-bd91-ce994043438f\") " pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.029526 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvmr9\" (UniqueName: \"kubernetes.io/projected/480a1412-80ec-4d41-bd91-ce994043438f-kube-api-access-zvmr9\") pod \"mysqld-exporter-91cb-account-create-update-7fmk8\" (UID: \"480a1412-80ec-4d41-bd91-ce994043438f\") " pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.030418 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/480a1412-80ec-4d41-bd91-ce994043438f-operator-scripts\") pod \"mysqld-exporter-91cb-account-create-update-7fmk8\" (UID: \"480a1412-80ec-4d41-bd91-ce994043438f\") " pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.058517 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvmr9\" (UniqueName: \"kubernetes.io/projected/480a1412-80ec-4d41-bd91-ce994043438f-kube-api-access-zvmr9\") pod \"mysqld-exporter-91cb-account-create-update-7fmk8\" (UID: \"480a1412-80ec-4d41-bd91-ce994043438f\") " pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.116671 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.196226 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lphwt"] Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.196983 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-lphwt" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerName="dnsmasq-dns" containerID="cri-o://dd17c1a35e8eb97fc1b40aee9145526c2e65db13c38cd7105d30888849a8ab1b" gracePeriod=10 Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.198486 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.246472 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.311843 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.368659 4912 generic.go:334] "Generic (PLEG): container finished" podID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerID="dd17c1a35e8eb97fc1b40aee9145526c2e65db13c38cd7105d30888849a8ab1b" exitCode=0 Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.368772 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lphwt" event={"ID":"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba","Type":"ContainerDied","Data":"dd17c1a35e8eb97fc1b40aee9145526c2e65db13c38cd7105d30888849a8ab1b"} Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.371388 4912 generic.go:334] "Generic (PLEG): container finished" podID="a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd" containerID="858cd40b2fa60b78d79641c0e6072a9ed2ff2e74af9da2047bee6ac5253aa75f" exitCode=0 Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.371483 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-s88vh" event={"ID":"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd","Type":"ContainerDied","Data":"858cd40b2fa60b78d79641c0e6072a9ed2ff2e74af9da2047bee6ac5253aa75f"} Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.379959 4912 generic.go:334] "Generic (PLEG): container finished" podID="ec53e2fd-3f5c-4382-a815-5e92e3299249" containerID="3e822a25c7ff8f47757b89553f4fe0ee778f9cc74ed8f6cdf5a8df204da4f123" exitCode=0 Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.380075 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-51b0-account-create-update-vqt9z" event={"ID":"ec53e2fd-3f5c-4382-a815-5e92e3299249","Type":"ContainerDied","Data":"3e822a25c7ff8f47757b89553f4fe0ee778f9cc74ed8f6cdf5a8df204da4f123"} Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.380120 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-51b0-account-create-update-vqt9z" event={"ID":"ec53e2fd-3f5c-4382-a815-5e92e3299249","Type":"ContainerStarted","Data":"88905b0fe0bc158f36b00282243c9403cdf74ff0697e10a04c2091b5bf8c38e2"} Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.421853 4912 generic.go:334] "Generic (PLEG): container finished" podID="5a43da64-c71a-4ae0-af3a-bc0fb42538ee" containerID="18732f729a350106351f369c1cb8ccc7f8a10f3bfc3f4f547aafa0b3d0486b83" exitCode=0 Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.421966 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1612-account-create-update-qg8wc" event={"ID":"5a43da64-c71a-4ae0-af3a-bc0fb42538ee","Type":"ContainerDied","Data":"18732f729a350106351f369c1cb8ccc7f8a10f3bfc3f4f547aafa0b3d0486b83"} Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.445855 4912 generic.go:334] "Generic (PLEG): container finished" podID="8e4aa219-c983-4ed9-96e6-e6583d25d34d" containerID="86093197118c3d966d8793f772f8b048f69db99608dbd3a11d170f362a81911c" exitCode=0 Dec 03 00:47:23 crc kubenswrapper[4912]: I1203 00:47:23.445932 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8dng2" event={"ID":"8e4aa219-c983-4ed9-96e6-e6583d25d34d","Type":"ContainerDied","Data":"86093197118c3d966d8793f772f8b048f69db99608dbd3a11d170f362a81911c"} Dec 03 00:47:25 crc kubenswrapper[4912]: I1203 00:47:25.841625 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:47:25 crc kubenswrapper[4912]: I1203 00:47:25.901142 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.091802 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4smn9"] Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.288350 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-4dvfk"] Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.290229 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.302299 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4dvfk"] Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.419652 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-fe58-account-create-update-l5dwm"] Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.421263 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.433324 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.438147 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nzjq\" (UniqueName: \"kubernetes.io/projected/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-kube-api-access-2nzjq\") pod \"glance-db-create-4dvfk\" (UID: \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\") " pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.438314 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-operator-scripts\") pod \"glance-db-create-4dvfk\" (UID: \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\") " pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.511277 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fe58-account-create-update-l5dwm"] Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.541169 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nzjq\" (UniqueName: \"kubernetes.io/projected/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-kube-api-access-2nzjq\") pod \"glance-db-create-4dvfk\" (UID: \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\") " pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.541267 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-operator-scripts\") pod \"glance-db-create-4dvfk\" (UID: \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\") " pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.541327 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15829b7d-eed7-425e-bd19-c2e496495444-operator-scripts\") pod \"glance-fe58-account-create-update-l5dwm\" (UID: \"15829b7d-eed7-425e-bd19-c2e496495444\") " pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.541410 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd77d\" (UniqueName: \"kubernetes.io/projected/15829b7d-eed7-425e-bd19-c2e496495444-kube-api-access-fd77d\") pod \"glance-fe58-account-create-update-l5dwm\" (UID: \"15829b7d-eed7-425e-bd19-c2e496495444\") " pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.542601 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-operator-scripts\") pod \"glance-db-create-4dvfk\" (UID: \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\") " pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.614297 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nzjq\" (UniqueName: \"kubernetes.io/projected/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-kube-api-access-2nzjq\") pod \"glance-db-create-4dvfk\" (UID: \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\") " pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.643229 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15829b7d-eed7-425e-bd19-c2e496495444-operator-scripts\") pod \"glance-fe58-account-create-update-l5dwm\" (UID: \"15829b7d-eed7-425e-bd19-c2e496495444\") " pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.643351 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd77d\" (UniqueName: \"kubernetes.io/projected/15829b7d-eed7-425e-bd19-c2e496495444-kube-api-access-fd77d\") pod \"glance-fe58-account-create-update-l5dwm\" (UID: \"15829b7d-eed7-425e-bd19-c2e496495444\") " pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.645108 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15829b7d-eed7-425e-bd19-c2e496495444-operator-scripts\") pod \"glance-fe58-account-create-update-l5dwm\" (UID: \"15829b7d-eed7-425e-bd19-c2e496495444\") " pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.645193 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.671150 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd77d\" (UniqueName: \"kubernetes.io/projected/15829b7d-eed7-425e-bd19-c2e496495444-kube-api-access-fd77d\") pod \"glance-fe58-account-create-update-l5dwm\" (UID: \"15829b7d-eed7-425e-bd19-c2e496495444\") " pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.751990 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:26 crc kubenswrapper[4912]: I1203 00:47:26.916823 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-lphwt" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: connect: connection refused" Dec 03 00:47:27 crc kubenswrapper[4912]: I1203 00:47:27.508765 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4smn9" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerName="registry-server" containerID="cri-o://bcfc7bca6c4259cc09d7cbb6628dba16e94f746d8e12e4e611ccfd15d71d4f36" gracePeriod=2 Dec 03 00:47:28 crc kubenswrapper[4912]: I1203 00:47:28.529320 4912 generic.go:334] "Generic (PLEG): container finished" podID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerID="bcfc7bca6c4259cc09d7cbb6628dba16e94f746d8e12e4e611ccfd15d71d4f36" exitCode=0 Dec 03 00:47:28 crc kubenswrapper[4912]: I1203 00:47:28.529375 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4smn9" event={"ID":"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f","Type":"ContainerDied","Data":"bcfc7bca6c4259cc09d7cbb6628dba16e94f746d8e12e4e611ccfd15d71d4f36"} Dec 03 00:47:29 crc kubenswrapper[4912]: I1203 00:47:29.829451 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:29 crc kubenswrapper[4912]: I1203 00:47:29.932163 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:29 crc kubenswrapper[4912]: E1203 00:47:29.932638 4912 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 00:47:29 crc kubenswrapper[4912]: E1203 00:47:29.932659 4912 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 00:47:29 crc kubenswrapper[4912]: E1203 00:47:29.932853 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift podName:761ebfbd-144d-42cc-ac4f-09a0c49ec7bb nodeName:}" failed. No retries permitted until 2025-12-03 00:47:45.932829683 +0000 UTC m=+1451.574850243 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift") pod "swift-storage-0" (UID: "761ebfbd-144d-42cc-ac4f-09a0c49ec7bb") : configmap "swift-ring-files" not found Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.037392 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9qg2\" (UniqueName: \"kubernetes.io/projected/8e4aa219-c983-4ed9-96e6-e6583d25d34d-kube-api-access-d9qg2\") pod \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\" (UID: \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\") " Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.037487 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e4aa219-c983-4ed9-96e6-e6583d25d34d-operator-scripts\") pod \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\" (UID: \"8e4aa219-c983-4ed9-96e6-e6583d25d34d\") " Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.038895 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e4aa219-c983-4ed9-96e6-e6583d25d34d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8e4aa219-c983-4ed9-96e6-e6583d25d34d" (UID: "8e4aa219-c983-4ed9-96e6-e6583d25d34d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.046676 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e4aa219-c983-4ed9-96e6-e6583d25d34d-kube-api-access-d9qg2" (OuterVolumeSpecName: "kube-api-access-d9qg2") pod "8e4aa219-c983-4ed9-96e6-e6583d25d34d" (UID: "8e4aa219-c983-4ed9-96e6-e6583d25d34d"). InnerVolumeSpecName "kube-api-access-d9qg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.140969 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9qg2\" (UniqueName: \"kubernetes.io/projected/8e4aa219-c983-4ed9-96e6-e6583d25d34d-kube-api-access-d9qg2\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.141020 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e4aa219-c983-4ed9-96e6-e6583d25d34d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.359744 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-5674795f8c-jxb67" podUID="bcee1be8-8dde-47a2-85ec-69af93765fe9" containerName="console" containerID="cri-o://abc6e08f599050799a8a4bad099d69cc8053722355adf6421ba2cea719bf3a84" gracePeriod=15 Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.636659 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5674795f8c-jxb67_bcee1be8-8dde-47a2-85ec-69af93765fe9/console/0.log" Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.637180 4912 generic.go:334] "Generic (PLEG): container finished" podID="bcee1be8-8dde-47a2-85ec-69af93765fe9" containerID="abc6e08f599050799a8a4bad099d69cc8053722355adf6421ba2cea719bf3a84" exitCode=2 Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.637271 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5674795f8c-jxb67" event={"ID":"bcee1be8-8dde-47a2-85ec-69af93765fe9","Type":"ContainerDied","Data":"abc6e08f599050799a8a4bad099d69cc8053722355adf6421ba2cea719bf3a84"} Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.639775 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-8dng2" event={"ID":"8e4aa219-c983-4ed9-96e6-e6583d25d34d","Type":"ContainerDied","Data":"ed96cd4a41334ddcb073b560b56f7f0e177e7778a628d519295e9798d5dbe3b3"} Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.639819 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed96cd4a41334ddcb073b560b56f7f0e177e7778a628d519295e9798d5dbe3b3" Dec 03 00:47:30 crc kubenswrapper[4912]: I1203 00:47:30.639837 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-8dng2" Dec 03 00:47:31 crc kubenswrapper[4912]: I1203 00:47:31.088464 4912 patch_prober.go:28] interesting pod/console-5674795f8c-jxb67 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.79:8443/health\": dial tcp 10.217.0.79:8443: connect: connection refused" start-of-body= Dec 03 00:47:31 crc kubenswrapper[4912]: I1203 00:47:31.088557 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-5674795f8c-jxb67" podUID="bcee1be8-8dde-47a2-85ec-69af93765fe9" containerName="console" probeResult="failure" output="Get \"https://10.217.0.79:8443/health\": dial tcp 10.217.0.79:8443: connect: connection refused" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.055516 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.122161 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:32 crc kubenswrapper[4912]: W1203 00:47:32.128704 4912 container.go:586] Failed to update stats for container "/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod197a0bff_4ebc_4581_bc0f_77d459500339.slice/crio-3b4472365e6ef87c7f98ea5a19b918bb41b441565289c7ebe2c78c33ccf8cb49": error while statting cgroup v2: [unable to parse /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod197a0bff_4ebc_4581_bc0f_77d459500339.slice/crio-3b4472365e6ef87c7f98ea5a19b918bb41b441565289c7ebe2c78c33ccf8cb49/memory.stat: read /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod197a0bff_4ebc_4581_bc0f_77d459500339.slice/crio-3b4472365e6ef87c7f98ea5a19b918bb41b441565289c7ebe2c78c33ccf8cb49/memory.stat: no such device], continuing to push stats Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.164513 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-s88vh" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.200353 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.247223 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpksk\" (UniqueName: \"kubernetes.io/projected/ec53e2fd-3f5c-4382-a815-5e92e3299249-kube-api-access-xpksk\") pod \"ec53e2fd-3f5c-4382-a815-5e92e3299249\" (UID: \"ec53e2fd-3f5c-4382-a815-5e92e3299249\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.247680 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp77p\" (UniqueName: \"kubernetes.io/projected/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-kube-api-access-vp77p\") pod \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\" (UID: \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.247755 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-operator-scripts\") pod \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\" (UID: \"5a43da64-c71a-4ae0-af3a-bc0fb42538ee\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.247806 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfv5z\" (UniqueName: \"kubernetes.io/projected/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-kube-api-access-jfv5z\") pod \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\" (UID: \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.247903 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdzv5\" (UniqueName: \"kubernetes.io/projected/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-kube-api-access-jdzv5\") pod \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.247934 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-utilities\") pod \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.247973 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-operator-scripts\") pod \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\" (UID: \"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.247992 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-catalog-content\") pod \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\" (UID: \"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.248011 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53e2fd-3f5c-4382-a815-5e92e3299249-operator-scripts\") pod \"ec53e2fd-3f5c-4382-a815-5e92e3299249\" (UID: \"ec53e2fd-3f5c-4382-a815-5e92e3299249\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.252263 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-utilities" (OuterVolumeSpecName: "utilities") pod "ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" (UID: "ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.253275 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd" (UID: "a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.253891 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a43da64-c71a-4ae0-af3a-bc0fb42538ee" (UID: "5a43da64-c71a-4ae0-af3a-bc0fb42538ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.253910 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec53e2fd-3f5c-4382-a815-5e92e3299249-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec53e2fd-3f5c-4382-a815-5e92e3299249" (UID: "ec53e2fd-3f5c-4382-a815-5e92e3299249"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.258612 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.279925 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-kube-api-access-jdzv5" (OuterVolumeSpecName: "kube-api-access-jdzv5") pod "ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" (UID: "ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f"). InnerVolumeSpecName "kube-api-access-jdzv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.285410 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec53e2fd-3f5c-4382-a815-5e92e3299249-kube-api-access-xpksk" (OuterVolumeSpecName: "kube-api-access-xpksk") pod "ec53e2fd-3f5c-4382-a815-5e92e3299249" (UID: "ec53e2fd-3f5c-4382-a815-5e92e3299249"). InnerVolumeSpecName "kube-api-access-xpksk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.286386 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-kube-api-access-vp77p" (OuterVolumeSpecName: "kube-api-access-vp77p") pod "5a43da64-c71a-4ae0-af3a-bc0fb42538ee" (UID: "5a43da64-c71a-4ae0-af3a-bc0fb42538ee"). InnerVolumeSpecName "kube-api-access-vp77p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.292510 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-kube-api-access-jfv5z" (OuterVolumeSpecName: "kube-api-access-jfv5z") pod "a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd" (UID: "a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd"). InnerVolumeSpecName "kube-api-access-jfv5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.328752 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5674795f8c-jxb67_bcee1be8-8dde-47a2-85ec-69af93765fe9/console/0.log" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.328818 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.359597 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l586b\" (UniqueName: \"kubernetes.io/projected/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-kube-api-access-l586b\") pod \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.359653 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-config\") pod \"bcee1be8-8dde-47a2-85ec-69af93765fe9\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.359696 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-service-ca\") pod \"bcee1be8-8dde-47a2-85ec-69af93765fe9\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.359814 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-config\") pod \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.359875 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2s6l\" (UniqueName: \"kubernetes.io/projected/bcee1be8-8dde-47a2-85ec-69af93765fe9-kube-api-access-q2s6l\") pod \"bcee1be8-8dde-47a2-85ec-69af93765fe9\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.359906 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-sb\") pod \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.363505 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-dns-svc\") pod \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.363562 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-nb\") pod \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\" (UID: \"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.363708 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-oauth-config\") pod \"bcee1be8-8dde-47a2-85ec-69af93765fe9\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.363766 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-trusted-ca-bundle\") pod \"bcee1be8-8dde-47a2-85ec-69af93765fe9\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.364815 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-oauth-serving-cert\") pod \"bcee1be8-8dde-47a2-85ec-69af93765fe9\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.365024 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-serving-cert\") pod \"bcee1be8-8dde-47a2-85ec-69af93765fe9\" (UID: \"bcee1be8-8dde-47a2-85ec-69af93765fe9\") " Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.366120 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.366143 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec53e2fd-3f5c-4382-a815-5e92e3299249-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.366158 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpksk\" (UniqueName: \"kubernetes.io/projected/ec53e2fd-3f5c-4382-a815-5e92e3299249-kube-api-access-xpksk\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.366172 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp77p\" (UniqueName: \"kubernetes.io/projected/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-kube-api-access-vp77p\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.366189 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a43da64-c71a-4ae0-af3a-bc0fb42538ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.366203 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfv5z\" (UniqueName: \"kubernetes.io/projected/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd-kube-api-access-jfv5z\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.366219 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdzv5\" (UniqueName: \"kubernetes.io/projected/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-kube-api-access-jdzv5\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.366233 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.370921 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-kube-api-access-l586b" (OuterVolumeSpecName: "kube-api-access-l586b") pod "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" (UID: "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba"). InnerVolumeSpecName "kube-api-access-l586b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.371311 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "bcee1be8-8dde-47a2-85ec-69af93765fe9" (UID: "bcee1be8-8dde-47a2-85ec-69af93765fe9"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.371361 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-config" (OuterVolumeSpecName: "console-config") pod "bcee1be8-8dde-47a2-85ec-69af93765fe9" (UID: "bcee1be8-8dde-47a2-85ec-69af93765fe9"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.372075 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "bcee1be8-8dde-47a2-85ec-69af93765fe9" (UID: "bcee1be8-8dde-47a2-85ec-69af93765fe9"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.372549 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-service-ca" (OuterVolumeSpecName: "service-ca") pod "bcee1be8-8dde-47a2-85ec-69af93765fe9" (UID: "bcee1be8-8dde-47a2-85ec-69af93765fe9"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.372634 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "bcee1be8-8dde-47a2-85ec-69af93765fe9" (UID: "bcee1be8-8dde-47a2-85ec-69af93765fe9"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.388102 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "bcee1be8-8dde-47a2-85ec-69af93765fe9" (UID: "bcee1be8-8dde-47a2-85ec-69af93765fe9"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.389831 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcee1be8-8dde-47a2-85ec-69af93765fe9-kube-api-access-q2s6l" (OuterVolumeSpecName: "kube-api-access-q2s6l") pod "bcee1be8-8dde-47a2-85ec-69af93765fe9" (UID: "bcee1be8-8dde-47a2-85ec-69af93765fe9"). InnerVolumeSpecName "kube-api-access-q2s6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.425267 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" (UID: "ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.433346 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-config" (OuterVolumeSpecName: "config") pod "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" (UID: "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.440977 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" (UID: "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.442422 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" (UID: "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.442559 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" (UID: "0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.468245 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.468996 4912 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469014 4912 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469031 4912 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469044 4912 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469089 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l586b\" (UniqueName: \"kubernetes.io/projected/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-kube-api-access-l586b\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469108 4912 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469119 4912 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bcee1be8-8dde-47a2-85ec-69af93765fe9-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469163 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469177 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2s6l\" (UniqueName: \"kubernetes.io/projected/bcee1be8-8dde-47a2-85ec-69af93765fe9-kube-api-access-q2s6l\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469192 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469204 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.469246 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.667403 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5674795f8c-jxb67_bcee1be8-8dde-47a2-85ec-69af93765fe9/console/0.log" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.667582 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5674795f8c-jxb67" event={"ID":"bcee1be8-8dde-47a2-85ec-69af93765fe9","Type":"ContainerDied","Data":"2c3d574f42cdf24e131647b1cff087eb2e8a7e58bf26e15ce78cfd28b45f3b26"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.667609 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5674795f8c-jxb67" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.667677 4912 scope.go:117] "RemoveContainer" containerID="abc6e08f599050799a8a4bad099d69cc8053722355adf6421ba2cea719bf3a84" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.673444 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4smn9" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.674397 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4smn9" event={"ID":"ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f","Type":"ContainerDied","Data":"d28a8329f9eb3f64a1e9c57aaac6b84d983182e4f548711e18f84f57d919e1ea"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.688035 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerStarted","Data":"d62cd23fa7798120cbc7fbf3e89799268dfd494b1f4854490741343fa74ec4f8"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.690089 4912 generic.go:334] "Generic (PLEG): container finished" podID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" containerID="d601bbaa0a1e7902f9bd53059640462774984a5b945e216fc8163f7c3ae23a3c" exitCode=0 Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.690173 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2","Type":"ContainerDied","Data":"d601bbaa0a1e7902f9bd53059640462774984a5b945e216fc8163f7c3ae23a3c"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.692785 4912 scope.go:117] "RemoveContainer" containerID="bcfc7bca6c4259cc09d7cbb6628dba16e94f746d8e12e4e611ccfd15d71d4f36" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.697533 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6wv8k" event={"ID":"1556422b-130c-4bf1-8010-ee7c6ea5f662","Type":"ContainerStarted","Data":"4a6214177ebb442eecf1dea7178ad692882345273c20d78e8889908d4e688268"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.729378 4912 generic.go:334] "Generic (PLEG): container finished" podID="38bf4c96-49c7-432f-be55-159b75ed1243" containerID="dff02ed2edeeec80c4c16589ce57d80d6f4bff8f20597991b6274a1f56ab981a" exitCode=0 Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.729468 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"38bf4c96-49c7-432f-be55-159b75ed1243","Type":"ContainerDied","Data":"dff02ed2edeeec80c4c16589ce57d80d6f4bff8f20597991b6274a1f56ab981a"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.738405 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5674795f8c-jxb67"] Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.752983 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-1612-account-create-update-qg8wc" event={"ID":"5a43da64-c71a-4ae0-af3a-bc0fb42538ee","Type":"ContainerDied","Data":"9d4d42bb18bbe64ed602b9de7e877100b774603d9a50be259d494cca77ddcb02"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.753047 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d4d42bb18bbe64ed602b9de7e877100b774603d9a50be259d494cca77ddcb02" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.753195 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-1612-account-create-update-qg8wc" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.765397 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"197a0bff-4ebc-4581-bc0f-77d459500339","Type":"ContainerStarted","Data":"4df618bd0cf191e58200c67750cfa36490cdd6bdf58990c3138aea64a555eec3"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.765539 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"197a0bff-4ebc-4581-bc0f-77d459500339","Type":"ContainerStarted","Data":"5b1b5f9bd52b982666368b538e1a71e14476c8bee43eaa7fb2044cb39f63f878"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.765715 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.772815 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5674795f8c-jxb67"] Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.777286 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lphwt" event={"ID":"0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba","Type":"ContainerDied","Data":"a247c66dac65b4ee2675b27955080f648d953fd1c63fa7a6c00aab0aa083ee9a"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.777423 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-lphwt" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.783884 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4smn9"] Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.794304 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-s88vh" event={"ID":"a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd","Type":"ContainerDied","Data":"c67d8792e77507b7665e93ce63c51cd71fc7662e6a36f5c31a65d95aa4b31b29"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.794368 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c67d8792e77507b7665e93ce63c51cd71fc7662e6a36f5c31a65d95aa4b31b29" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.794895 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4smn9"] Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.796509 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-s88vh" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.796961 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-6wv8k" podStartSLOduration=2.700379227 podStartE2EDuration="15.796938402s" podCreationTimestamp="2025-12-03 00:47:17 +0000 UTC" firstStartedPulling="2025-12-03 00:47:18.753204882 +0000 UTC m=+1424.395225442" lastFinishedPulling="2025-12-03 00:47:31.849764057 +0000 UTC m=+1437.491784617" observedRunningTime="2025-12-03 00:47:32.745206371 +0000 UTC m=+1438.387226981" watchObservedRunningTime="2025-12-03 00:47:32.796938402 +0000 UTC m=+1438.438958962" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.811724 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-51b0-account-create-update-vqt9z" event={"ID":"ec53e2fd-3f5c-4382-a815-5e92e3299249","Type":"ContainerDied","Data":"88905b0fe0bc158f36b00282243c9403cdf74ff0697e10a04c2091b5bf8c38e2"} Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.811766 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88905b0fe0bc158f36b00282243c9403cdf74ff0697e10a04c2091b5bf8c38e2" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.811873 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-51b0-account-create-update-vqt9z" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.819744 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.609720302 podStartE2EDuration="13.819720841s" podCreationTimestamp="2025-12-03 00:47:19 +0000 UTC" firstStartedPulling="2025-12-03 00:47:20.68516111 +0000 UTC m=+1426.327181670" lastFinishedPulling="2025-12-03 00:47:31.895161649 +0000 UTC m=+1437.537182209" observedRunningTime="2025-12-03 00:47:32.808636765 +0000 UTC m=+1438.450657335" watchObservedRunningTime="2025-12-03 00:47:32.819720841 +0000 UTC m=+1438.461741401" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.944322 4912 scope.go:117] "RemoveContainer" containerID="3f58a122668d0e2ab731a664ea789afbb31c750daa6f21bdb5d0ad54907a9ffa" Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.972774 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lphwt"] Dec 03 00:47:32 crc kubenswrapper[4912]: I1203 00:47:32.988029 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lphwt"] Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.030828 4912 scope.go:117] "RemoveContainer" containerID="5f5522f339ab2c5d4dc14dc9af58e23cdb8b198cf2ce18b2e927cc23c4b15657" Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.108853 4912 scope.go:117] "RemoveContainer" containerID="dd17c1a35e8eb97fc1b40aee9145526c2e65db13c38cd7105d30888849a8ab1b" Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.138182 4912 scope.go:117] "RemoveContainer" containerID="e09847018f67cff28ea0a2e1b264c159721c31ed141fc0a7d26cc92f37cd7639" Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.366521 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-htd7z"] Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.389796 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-4dvfk"] Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.416018 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-91cb-account-create-update-7fmk8"] Dec 03 00:47:33 crc kubenswrapper[4912]: W1203 00:47:33.419742 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod480a1412_80ec_4d41_bd91_ce994043438f.slice/crio-41a240b579a057a4ba6d3e162ab1735e730152ee804ef5c7afde0575dd56f373 WatchSource:0}: Error finding container 41a240b579a057a4ba6d3e162ab1735e730152ee804ef5c7afde0575dd56f373: Status 404 returned error can't find the container with id 41a240b579a057a4ba6d3e162ab1735e730152ee804ef5c7afde0575dd56f373 Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.463945 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fe58-account-create-update-l5dwm"] Dec 03 00:47:33 crc kubenswrapper[4912]: W1203 00:47:33.483162 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15829b7d_eed7_425e_bd19_c2e496495444.slice/crio-2367fe6ea6e83a2594fda83a2a246aa794311eb9fb5e4f2824a47a7c3f9ae3c9 WatchSource:0}: Error finding container 2367fe6ea6e83a2594fda83a2a246aa794311eb9fb5e4f2824a47a7c3f9ae3c9: Status 404 returned error can't find the container with id 2367fe6ea6e83a2594fda83a2a246aa794311eb9fb5e4f2824a47a7c3f9ae3c9 Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.825731 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2","Type":"ContainerStarted","Data":"f4b259a28c47376f179f856aa118415fa4f7ca51f67b049df6e01b70110ca910"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.826365 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.829766 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-htd7z" event={"ID":"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598","Type":"ContainerStarted","Data":"b89ae9f2492911940bbbafaa992074af07db0735db2f13da75c218796efcd454"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.829832 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-htd7z" event={"ID":"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598","Type":"ContainerStarted","Data":"1699a6c0097bdba27460672ec534444ca42e35e761690b18ca26915da2061db2"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.833615 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"38bf4c96-49c7-432f-be55-159b75ed1243","Type":"ContainerStarted","Data":"c303a0e01c5dd3aed0f794b31d152ce2a5a78a187efb125ebddda24df8be9ae5"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.833928 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.837654 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4dvfk" event={"ID":"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03","Type":"ContainerStarted","Data":"f27b4b82e446bd02a99dc89d09ea3966c4dd87335a069a2080d9eb3d4f20a933"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.837693 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4dvfk" event={"ID":"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03","Type":"ContainerStarted","Data":"3edba83388149a124190e71b78f2cb4256d44021b468011dd942b79001b169a9"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.842682 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" event={"ID":"480a1412-80ec-4d41-bd91-ce994043438f","Type":"ContainerStarted","Data":"dca309568638bb16bf1be681679992b85fff17331db10a4e2d5df45e0a0b1699"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.842749 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" event={"ID":"480a1412-80ec-4d41-bd91-ce994043438f","Type":"ContainerStarted","Data":"41a240b579a057a4ba6d3e162ab1735e730152ee804ef5c7afde0575dd56f373"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.845093 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fe58-account-create-update-l5dwm" event={"ID":"15829b7d-eed7-425e-bd19-c2e496495444","Type":"ContainerStarted","Data":"097553effd16008a2285a5950f01f07946d1046482f3611ac7b810bfecfc9bd3"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.845141 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fe58-account-create-update-l5dwm" event={"ID":"15829b7d-eed7-425e-bd19-c2e496495444","Type":"ContainerStarted","Data":"2367fe6ea6e83a2594fda83a2a246aa794311eb9fb5e4f2824a47a7c3f9ae3c9"} Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.862918 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.583340112 podStartE2EDuration="1m17.862894711s" podCreationTimestamp="2025-12-03 00:46:16 +0000 UTC" firstStartedPulling="2025-12-03 00:46:18.978743574 +0000 UTC m=+1364.620764144" lastFinishedPulling="2025-12-03 00:46:57.258298183 +0000 UTC m=+1402.900318743" observedRunningTime="2025-12-03 00:47:33.854921988 +0000 UTC m=+1439.496942548" watchObservedRunningTime="2025-12-03 00:47:33.862894711 +0000 UTC m=+1439.504915271" Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.895235 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-db-create-htd7z" podStartSLOduration=11.895214383999999 podStartE2EDuration="11.895214384s" podCreationTimestamp="2025-12-03 00:47:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:33.888487854 +0000 UTC m=+1439.530508424" watchObservedRunningTime="2025-12-03 00:47:33.895214384 +0000 UTC m=+1439.537234944" Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.909946 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-4dvfk" podStartSLOduration=7.909922546 podStartE2EDuration="7.909922546s" podCreationTimestamp="2025-12-03 00:47:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:33.907312696 +0000 UTC m=+1439.549333266" watchObservedRunningTime="2025-12-03 00:47:33.909922546 +0000 UTC m=+1439.551943106" Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.946912 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-fe58-account-create-update-l5dwm" podStartSLOduration=7.946889764 podStartE2EDuration="7.946889764s" podCreationTimestamp="2025-12-03 00:47:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:33.932914521 +0000 UTC m=+1439.574935081" watchObservedRunningTime="2025-12-03 00:47:33.946889764 +0000 UTC m=+1439.588910324" Dec 03 00:47:33 crc kubenswrapper[4912]: I1203 00:47:33.981926 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" podStartSLOduration=11.981904109 podStartE2EDuration="11.981904109s" podCreationTimestamp="2025-12-03 00:47:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:33.975835167 +0000 UTC m=+1439.617855727" watchObservedRunningTime="2025-12-03 00:47:33.981904109 +0000 UTC m=+1439.623924669" Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.058715 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.5186476 podStartE2EDuration="1m18.058684089s" podCreationTimestamp="2025-12-03 00:46:16 +0000 UTC" firstStartedPulling="2025-12-03 00:46:18.718567052 +0000 UTC m=+1364.360587612" lastFinishedPulling="2025-12-03 00:46:57.258603541 +0000 UTC m=+1402.900624101" observedRunningTime="2025-12-03 00:47:34.054468156 +0000 UTC m=+1439.696488726" watchObservedRunningTime="2025-12-03 00:47:34.058684089 +0000 UTC m=+1439.700704649" Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.592755 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" path="/var/lib/kubelet/pods/0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba/volumes" Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.593536 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" path="/var/lib/kubelet/pods/ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f/volumes" Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.594527 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcee1be8-8dde-47a2-85ec-69af93765fe9" path="/var/lib/kubelet/pods/bcee1be8-8dde-47a2-85ec-69af93765fe9/volumes" Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.886728 4912 generic.go:334] "Generic (PLEG): container finished" podID="eeaa6d61-6f8d-44dd-bd9b-708c59ab6598" containerID="b89ae9f2492911940bbbafaa992074af07db0735db2f13da75c218796efcd454" exitCode=0 Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.887503 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-htd7z" event={"ID":"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598","Type":"ContainerDied","Data":"b89ae9f2492911940bbbafaa992074af07db0735db2f13da75c218796efcd454"} Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.890824 4912 generic.go:334] "Generic (PLEG): container finished" podID="480a1412-80ec-4d41-bd91-ce994043438f" containerID="dca309568638bb16bf1be681679992b85fff17331db10a4e2d5df45e0a0b1699" exitCode=0 Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.890930 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" event={"ID":"480a1412-80ec-4d41-bd91-ce994043438f","Type":"ContainerDied","Data":"dca309568638bb16bf1be681679992b85fff17331db10a4e2d5df45e0a0b1699"} Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.898555 4912 generic.go:334] "Generic (PLEG): container finished" podID="15829b7d-eed7-425e-bd19-c2e496495444" containerID="097553effd16008a2285a5950f01f07946d1046482f3611ac7b810bfecfc9bd3" exitCode=0 Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.898654 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fe58-account-create-update-l5dwm" event={"ID":"15829b7d-eed7-425e-bd19-c2e496495444","Type":"ContainerDied","Data":"097553effd16008a2285a5950f01f07946d1046482f3611ac7b810bfecfc9bd3"} Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.905771 4912 generic.go:334] "Generic (PLEG): container finished" podID="3c8872b9-97a4-48d7-bf51-bab0b0a5cf03" containerID="f27b4b82e446bd02a99dc89d09ea3966c4dd87335a069a2080d9eb3d4f20a933" exitCode=0 Dec 03 00:47:34 crc kubenswrapper[4912]: I1203 00:47:34.905909 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4dvfk" event={"ID":"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03","Type":"ContainerDied","Data":"f27b4b82e446bd02a99dc89d09ea3966c4dd87335a069a2080d9eb3d4f20a933"} Dec 03 00:47:35 crc kubenswrapper[4912]: I1203 00:47:35.917862 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerStarted","Data":"0879a34d96d5a73e15cfbe99d2677a45352f81b80652556076f6cab36d64c53b"} Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.410397 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.489980 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvmr9\" (UniqueName: \"kubernetes.io/projected/480a1412-80ec-4d41-bd91-ce994043438f-kube-api-access-zvmr9\") pod \"480a1412-80ec-4d41-bd91-ce994043438f\" (UID: \"480a1412-80ec-4d41-bd91-ce994043438f\") " Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.490062 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/480a1412-80ec-4d41-bd91-ce994043438f-operator-scripts\") pod \"480a1412-80ec-4d41-bd91-ce994043438f\" (UID: \"480a1412-80ec-4d41-bd91-ce994043438f\") " Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.491558 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/480a1412-80ec-4d41-bd91-ce994043438f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "480a1412-80ec-4d41-bd91-ce994043438f" (UID: "480a1412-80ec-4d41-bd91-ce994043438f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.511226 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480a1412-80ec-4d41-bd91-ce994043438f-kube-api-access-zvmr9" (OuterVolumeSpecName: "kube-api-access-zvmr9") pod "480a1412-80ec-4d41-bd91-ce994043438f" (UID: "480a1412-80ec-4d41-bd91-ce994043438f"). InnerVolumeSpecName "kube-api-access-zvmr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.599728 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.600206 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvmr9\" (UniqueName: \"kubernetes.io/projected/480a1412-80ec-4d41-bd91-ce994043438f-kube-api-access-zvmr9\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.600508 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/480a1412-80ec-4d41-bd91-ce994043438f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.609472 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.701361 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15829b7d-eed7-425e-bd19-c2e496495444-operator-scripts\") pod \"15829b7d-eed7-425e-bd19-c2e496495444\" (UID: \"15829b7d-eed7-425e-bd19-c2e496495444\") " Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.701649 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd77d\" (UniqueName: \"kubernetes.io/projected/15829b7d-eed7-425e-bd19-c2e496495444-kube-api-access-fd77d\") pod \"15829b7d-eed7-425e-bd19-c2e496495444\" (UID: \"15829b7d-eed7-425e-bd19-c2e496495444\") " Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.701732 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-operator-scripts\") pod \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\" (UID: \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\") " Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.701768 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nzjq\" (UniqueName: \"kubernetes.io/projected/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-kube-api-access-2nzjq\") pod \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\" (UID: \"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03\") " Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.703333 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15829b7d-eed7-425e-bd19-c2e496495444-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "15829b7d-eed7-425e-bd19-c2e496495444" (UID: "15829b7d-eed7-425e-bd19-c2e496495444"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.704395 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3c8872b9-97a4-48d7-bf51-bab0b0a5cf03" (UID: "3c8872b9-97a4-48d7-bf51-bab0b0a5cf03"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.708744 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-kube-api-access-2nzjq" (OuterVolumeSpecName: "kube-api-access-2nzjq") pod "3c8872b9-97a4-48d7-bf51-bab0b0a5cf03" (UID: "3c8872b9-97a4-48d7-bf51-bab0b0a5cf03"). InnerVolumeSpecName "kube-api-access-2nzjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.709008 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15829b7d-eed7-425e-bd19-c2e496495444-kube-api-access-fd77d" (OuterVolumeSpecName: "kube-api-access-fd77d") pod "15829b7d-eed7-425e-bd19-c2e496495444" (UID: "15829b7d-eed7-425e-bd19-c2e496495444"). InnerVolumeSpecName "kube-api-access-fd77d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.770823 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.804018 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd77d\" (UniqueName: \"kubernetes.io/projected/15829b7d-eed7-425e-bd19-c2e496495444-kube-api-access-fd77d\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.804061 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.804071 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nzjq\" (UniqueName: \"kubernetes.io/projected/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03-kube-api-access-2nzjq\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.804082 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15829b7d-eed7-425e-bd19-c2e496495444-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.905893 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-operator-scripts\") pod \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\" (UID: \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\") " Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.906044 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jm2m\" (UniqueName: \"kubernetes.io/projected/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-kube-api-access-4jm2m\") pod \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\" (UID: \"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598\") " Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.908205 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eeaa6d61-6f8d-44dd-bd9b-708c59ab6598" (UID: "eeaa6d61-6f8d-44dd-bd9b-708c59ab6598"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.910462 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-kube-api-access-4jm2m" (OuterVolumeSpecName: "kube-api-access-4jm2m") pod "eeaa6d61-6f8d-44dd-bd9b-708c59ab6598" (UID: "eeaa6d61-6f8d-44dd-bd9b-708c59ab6598"). InnerVolumeSpecName "kube-api-access-4jm2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.916501 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-lphwt" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: i/o timeout" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.928265 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fe58-account-create-update-l5dwm" event={"ID":"15829b7d-eed7-425e-bd19-c2e496495444","Type":"ContainerDied","Data":"2367fe6ea6e83a2594fda83a2a246aa794311eb9fb5e4f2824a47a7c3f9ae3c9"} Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.928314 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2367fe6ea6e83a2594fda83a2a246aa794311eb9fb5e4f2824a47a7c3f9ae3c9" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.928367 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fe58-account-create-update-l5dwm" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.931653 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-4dvfk" event={"ID":"3c8872b9-97a4-48d7-bf51-bab0b0a5cf03","Type":"ContainerDied","Data":"3edba83388149a124190e71b78f2cb4256d44021b468011dd942b79001b169a9"} Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.931705 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3edba83388149a124190e71b78f2cb4256d44021b468011dd942b79001b169a9" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.931717 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-4dvfk" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.933595 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-htd7z" event={"ID":"eeaa6d61-6f8d-44dd-bd9b-708c59ab6598","Type":"ContainerDied","Data":"1699a6c0097bdba27460672ec534444ca42e35e761690b18ca26915da2061db2"} Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.933642 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1699a6c0097bdba27460672ec534444ca42e35e761690b18ca26915da2061db2" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.933610 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-htd7z" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.935226 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" event={"ID":"480a1412-80ec-4d41-bd91-ce994043438f","Type":"ContainerDied","Data":"41a240b579a057a4ba6d3e162ab1735e730152ee804ef5c7afde0575dd56f373"} Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.935253 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41a240b579a057a4ba6d3e162ab1735e730152ee804ef5c7afde0575dd56f373" Dec 03 00:47:36 crc kubenswrapper[4912]: I1203 00:47:36.935284 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-91cb-account-create-update-7fmk8" Dec 03 00:47:37 crc kubenswrapper[4912]: I1203 00:47:37.009329 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jm2m\" (UniqueName: \"kubernetes.io/projected/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-kube-api-access-4jm2m\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:37 crc kubenswrapper[4912]: I1203 00:47:37.009377 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:37 crc kubenswrapper[4912]: I1203 00:47:37.602547 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gbbh7" podUID="dd4efef6-5ce3-444a-a464-1f18f7cc2db4" containerName="ovn-controller" probeResult="failure" output=< Dec 03 00:47:37 crc kubenswrapper[4912]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 00:47:37 crc kubenswrapper[4912]: > Dec 03 00:47:37 crc kubenswrapper[4912]: I1203 00:47:37.651280 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:47:37 crc kubenswrapper[4912]: I1203 00:47:37.706348 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6lm7j" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.100318 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gbbh7-config-q8574"] Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101099 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4aa219-c983-4ed9-96e6-e6583d25d34d" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101119 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4aa219-c983-4ed9-96e6-e6583d25d34d" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101151 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480a1412-80ec-4d41-bd91-ce994043438f" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101157 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="480a1412-80ec-4d41-bd91-ce994043438f" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101169 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcee1be8-8dde-47a2-85ec-69af93765fe9" containerName="console" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101176 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcee1be8-8dde-47a2-85ec-69af93765fe9" containerName="console" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101184 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerName="dnsmasq-dns" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101190 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerName="dnsmasq-dns" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101204 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101210 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101218 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerName="extract-utilities" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101224 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerName="extract-utilities" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101233 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec53e2fd-3f5c-4382-a815-5e92e3299249" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101239 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec53e2fd-3f5c-4382-a815-5e92e3299249" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101251 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerName="init" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101256 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerName="init" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101269 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15829b7d-eed7-425e-bd19-c2e496495444" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101278 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="15829b7d-eed7-425e-bd19-c2e496495444" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101289 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a43da64-c71a-4ae0-af3a-bc0fb42538ee" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101296 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a43da64-c71a-4ae0-af3a-bc0fb42538ee" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101305 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c8872b9-97a4-48d7-bf51-bab0b0a5cf03" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101311 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c8872b9-97a4-48d7-bf51-bab0b0a5cf03" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101325 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerName="extract-content" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101332 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerName="extract-content" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101345 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerName="registry-server" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101351 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerName="registry-server" Dec 03 00:47:38 crc kubenswrapper[4912]: E1203 00:47:38.101357 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeaa6d61-6f8d-44dd-bd9b-708c59ab6598" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101364 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeaa6d61-6f8d-44dd-bd9b-708c59ab6598" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101561 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="15829b7d-eed7-425e-bd19-c2e496495444" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101581 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="480a1412-80ec-4d41-bd91-ce994043438f" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101607 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c8872b9-97a4-48d7-bf51-bab0b0a5cf03" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101617 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a43da64-c71a-4ae0-af3a-bc0fb42538ee" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101628 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e4aa219-c983-4ed9-96e6-e6583d25d34d" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101637 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcee1be8-8dde-47a2-85ec-69af93765fe9" containerName="console" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101648 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101657 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d0c1d3b-5d5d-4d26-bc5b-3a6dddd5acba" containerName="dnsmasq-dns" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101667 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab125fcb-b1cf-48b0-a91d-2d9dfd7fe60f" containerName="registry-server" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101673 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeaa6d61-6f8d-44dd-bd9b-708c59ab6598" containerName="mariadb-database-create" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.101699 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec53e2fd-3f5c-4382-a815-5e92e3299249" containerName="mariadb-account-create-update" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.102690 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.109499 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.119462 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gbbh7-config-q8574"] Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.233919 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-log-ovn\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.233975 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.234009 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run-ovn\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.234102 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nq5w\" (UniqueName: \"kubernetes.io/projected/112fcd64-ec5f-492c-94ee-731a66de6b63-kube-api-access-9nq5w\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.234135 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-additional-scripts\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.234153 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-scripts\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.336560 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-log-ovn\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.336618 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.336649 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run-ovn\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.336935 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nq5w\" (UniqueName: \"kubernetes.io/projected/112fcd64-ec5f-492c-94ee-731a66de6b63-kube-api-access-9nq5w\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.336964 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-additional-scripts\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.336986 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-scripts\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.337230 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.337236 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run-ovn\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.337351 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-log-ovn\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.337843 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-additional-scripts\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.339289 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-scripts\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.358887 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nq5w\" (UniqueName: \"kubernetes.io/projected/112fcd64-ec5f-492c-94ee-731a66de6b63-kube-api-access-9nq5w\") pod \"ovn-controller-gbbh7-config-q8574\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:38 crc kubenswrapper[4912]: I1203 00:47:38.424968 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:39 crc kubenswrapper[4912]: I1203 00:47:39.987745 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerStarted","Data":"f99af3ce42b0c9f1459f5bc50c596e75148a7ed8d7be5a36b61b602acbda2801"} Dec 03 00:47:40 crc kubenswrapper[4912]: I1203 00:47:40.029716 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=37.078064249 podStartE2EDuration="1m18.029694454s" podCreationTimestamp="2025-12-03 00:46:22 +0000 UTC" firstStartedPulling="2025-12-03 00:46:58.6126489 +0000 UTC m=+1404.254669450" lastFinishedPulling="2025-12-03 00:47:39.564279085 +0000 UTC m=+1445.206299655" observedRunningTime="2025-12-03 00:47:40.028225885 +0000 UTC m=+1445.670246445" watchObservedRunningTime="2025-12-03 00:47:40.029694454 +0000 UTC m=+1445.671715014" Dec 03 00:47:40 crc kubenswrapper[4912]: I1203 00:47:40.097208 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gbbh7-config-q8574"] Dec 03 00:47:40 crc kubenswrapper[4912]: W1203 00:47:40.109666 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod112fcd64_ec5f_492c_94ee_731a66de6b63.slice/crio-78f78d2cb43e7a30dcb2aa9c56d799eb152d1e2d8e30e0c6079ee1afd966c067 WatchSource:0}: Error finding container 78f78d2cb43e7a30dcb2aa9c56d799eb152d1e2d8e30e0c6079ee1afd966c067: Status 404 returned error can't find the container with id 78f78d2cb43e7a30dcb2aa9c56d799eb152d1e2d8e30e0c6079ee1afd966c067 Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.002299 4912 generic.go:334] "Generic (PLEG): container finished" podID="112fcd64-ec5f-492c-94ee-731a66de6b63" containerID="02b42b411f8721761c9b9e86ca9993014177dc7bbcf2bcb5225128c8ff217eaf" exitCode=0 Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.002400 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gbbh7-config-q8574" event={"ID":"112fcd64-ec5f-492c-94ee-731a66de6b63","Type":"ContainerDied","Data":"02b42b411f8721761c9b9e86ca9993014177dc7bbcf2bcb5225128c8ff217eaf"} Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.003088 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gbbh7-config-q8574" event={"ID":"112fcd64-ec5f-492c-94ee-731a66de6b63","Type":"ContainerStarted","Data":"78f78d2cb43e7a30dcb2aa9c56d799eb152d1e2d8e30e0c6079ee1afd966c067"} Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.007176 4912 generic.go:334] "Generic (PLEG): container finished" podID="1556422b-130c-4bf1-8010-ee7c6ea5f662" containerID="4a6214177ebb442eecf1dea7178ad692882345273c20d78e8889908d4e688268" exitCode=0 Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.007273 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6wv8k" event={"ID":"1556422b-130c-4bf1-8010-ee7c6ea5f662","Type":"ContainerDied","Data":"4a6214177ebb442eecf1dea7178ad692882345273c20d78e8889908d4e688268"} Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.592083 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-tnx9g"] Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.593756 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.595901 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bvvth" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.600326 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.609610 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tnx9g"] Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.715557 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-combined-ca-bundle\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.715654 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk7cz\" (UniqueName: \"kubernetes.io/projected/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-kube-api-access-mk7cz\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.715940 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-config-data\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.716127 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-db-sync-config-data\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.818503 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-combined-ca-bundle\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.818563 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk7cz\" (UniqueName: \"kubernetes.io/projected/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-kube-api-access-mk7cz\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.818630 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-config-data\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.818678 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-db-sync-config-data\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.826061 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-db-sync-config-data\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.837346 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-config-data\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.837379 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-combined-ca-bundle\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.874111 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk7cz\" (UniqueName: \"kubernetes.io/projected/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-kube-api-access-mk7cz\") pod \"glance-db-sync-tnx9g\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:41 crc kubenswrapper[4912]: I1203 00:47:41.914623 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tnx9g" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.608702 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-gbbh7" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.663471 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.680552 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.739763 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1556422b-130c-4bf1-8010-ee7c6ea5f662-etc-swift\") pod \"1556422b-130c-4bf1-8010-ee7c6ea5f662\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.739963 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-scripts\") pod \"1556422b-130c-4bf1-8010-ee7c6ea5f662\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.739986 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-scripts\") pod \"112fcd64-ec5f-492c-94ee-731a66de6b63\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740026 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-additional-scripts\") pod \"112fcd64-ec5f-492c-94ee-731a66de6b63\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740114 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run-ovn\") pod \"112fcd64-ec5f-492c-94ee-731a66de6b63\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740180 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-log-ovn\") pod \"112fcd64-ec5f-492c-94ee-731a66de6b63\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740211 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkksb\" (UniqueName: \"kubernetes.io/projected/1556422b-130c-4bf1-8010-ee7c6ea5f662-kube-api-access-zkksb\") pod \"1556422b-130c-4bf1-8010-ee7c6ea5f662\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740237 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-ring-data-devices\") pod \"1556422b-130c-4bf1-8010-ee7c6ea5f662\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740256 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nq5w\" (UniqueName: \"kubernetes.io/projected/112fcd64-ec5f-492c-94ee-731a66de6b63-kube-api-access-9nq5w\") pod \"112fcd64-ec5f-492c-94ee-731a66de6b63\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740288 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-swiftconf\") pod \"1556422b-130c-4bf1-8010-ee7c6ea5f662\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740319 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-combined-ca-bundle\") pod \"1556422b-130c-4bf1-8010-ee7c6ea5f662\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740381 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run\") pod \"112fcd64-ec5f-492c-94ee-731a66de6b63\" (UID: \"112fcd64-ec5f-492c-94ee-731a66de6b63\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.740406 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-dispersionconf\") pod \"1556422b-130c-4bf1-8010-ee7c6ea5f662\" (UID: \"1556422b-130c-4bf1-8010-ee7c6ea5f662\") " Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.743459 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1556422b-130c-4bf1-8010-ee7c6ea5f662" (UID: "1556422b-130c-4bf1-8010-ee7c6ea5f662"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.746364 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-scripts" (OuterVolumeSpecName: "scripts") pod "112fcd64-ec5f-492c-94ee-731a66de6b63" (UID: "112fcd64-ec5f-492c-94ee-731a66de6b63"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.747076 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1556422b-130c-4bf1-8010-ee7c6ea5f662-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1556422b-130c-4bf1-8010-ee7c6ea5f662" (UID: "1556422b-130c-4bf1-8010-ee7c6ea5f662"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.750989 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "112fcd64-ec5f-492c-94ee-731a66de6b63" (UID: "112fcd64-ec5f-492c-94ee-731a66de6b63"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.751390 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "112fcd64-ec5f-492c-94ee-731a66de6b63" (UID: "112fcd64-ec5f-492c-94ee-731a66de6b63"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.751537 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run" (OuterVolumeSpecName: "var-run") pod "112fcd64-ec5f-492c-94ee-731a66de6b63" (UID: "112fcd64-ec5f-492c-94ee-731a66de6b63"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.751951 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "112fcd64-ec5f-492c-94ee-731a66de6b63" (UID: "112fcd64-ec5f-492c-94ee-731a66de6b63"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.756239 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/112fcd64-ec5f-492c-94ee-731a66de6b63-kube-api-access-9nq5w" (OuterVolumeSpecName: "kube-api-access-9nq5w") pod "112fcd64-ec5f-492c-94ee-731a66de6b63" (UID: "112fcd64-ec5f-492c-94ee-731a66de6b63"). InnerVolumeSpecName "kube-api-access-9nq5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.756989 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tnx9g"] Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.764853 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1556422b-130c-4bf1-8010-ee7c6ea5f662-kube-api-access-zkksb" (OuterVolumeSpecName: "kube-api-access-zkksb") pod "1556422b-130c-4bf1-8010-ee7c6ea5f662" (UID: "1556422b-130c-4bf1-8010-ee7c6ea5f662"). InnerVolumeSpecName "kube-api-access-zkksb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.769229 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1556422b-130c-4bf1-8010-ee7c6ea5f662" (UID: "1556422b-130c-4bf1-8010-ee7c6ea5f662"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: W1203 00:47:42.776678 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode760382f_ba1d_4d51_bdd0_2cd62315c4ae.slice/crio-c7fca989ce4af296a93885a099803984be7d9f5ef3dd171ba02d35bd74f284a5 WatchSource:0}: Error finding container c7fca989ce4af296a93885a099803984be7d9f5ef3dd171ba02d35bd74f284a5: Status 404 returned error can't find the container with id c7fca989ce4af296a93885a099803984be7d9f5ef3dd171ba02d35bd74f284a5 Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.803688 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-scripts" (OuterVolumeSpecName: "scripts") pod "1556422b-130c-4bf1-8010-ee7c6ea5f662" (UID: "1556422b-130c-4bf1-8010-ee7c6ea5f662"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.803941 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1556422b-130c-4bf1-8010-ee7c6ea5f662" (UID: "1556422b-130c-4bf1-8010-ee7c6ea5f662"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.815590 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1556422b-130c-4bf1-8010-ee7c6ea5f662" (UID: "1556422b-130c-4bf1-8010-ee7c6ea5f662"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843020 4912 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843333 4912 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1556422b-130c-4bf1-8010-ee7c6ea5f662-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843409 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843490 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843558 4912 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/112fcd64-ec5f-492c-94ee-731a66de6b63-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843616 4912 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843665 4912 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843740 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkksb\" (UniqueName: \"kubernetes.io/projected/1556422b-130c-4bf1-8010-ee7c6ea5f662-kube-api-access-zkksb\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843802 4912 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1556422b-130c-4bf1-8010-ee7c6ea5f662-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843859 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nq5w\" (UniqueName: \"kubernetes.io/projected/112fcd64-ec5f-492c-94ee-731a66de6b63-kube-api-access-9nq5w\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843907 4912 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.843971 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1556422b-130c-4bf1-8010-ee7c6ea5f662-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.844046 4912 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/112fcd64-ec5f-492c-94ee-731a66de6b63-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.982083 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-rx69z"] Dec 03 00:47:42 crc kubenswrapper[4912]: E1203 00:47:42.983122 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112fcd64-ec5f-492c-94ee-731a66de6b63" containerName="ovn-config" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.983214 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="112fcd64-ec5f-492c-94ee-731a66de6b63" containerName="ovn-config" Dec 03 00:47:42 crc kubenswrapper[4912]: E1203 00:47:42.983326 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1556422b-130c-4bf1-8010-ee7c6ea5f662" containerName="swift-ring-rebalance" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.983408 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1556422b-130c-4bf1-8010-ee7c6ea5f662" containerName="swift-ring-rebalance" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.983748 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="112fcd64-ec5f-492c-94ee-731a66de6b63" containerName="ovn-config" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.983825 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="1556422b-130c-4bf1-8010-ee7c6ea5f662" containerName="swift-ring-rebalance" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.984863 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:42 crc kubenswrapper[4912]: I1203 00:47:42.997574 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-rx69z"] Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.030675 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gbbh7-config-q8574" event={"ID":"112fcd64-ec5f-492c-94ee-731a66de6b63","Type":"ContainerDied","Data":"78f78d2cb43e7a30dcb2aa9c56d799eb152d1e2d8e30e0c6079ee1afd966c067"} Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.030930 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f78d2cb43e7a30dcb2aa9c56d799eb152d1e2d8e30e0c6079ee1afd966c067" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.030745 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gbbh7-config-q8574" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.032906 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tnx9g" event={"ID":"e760382f-ba1d-4d51-bdd0-2cd62315c4ae","Type":"ContainerStarted","Data":"c7fca989ce4af296a93885a099803984be7d9f5ef3dd171ba02d35bd74f284a5"} Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.034955 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6wv8k" event={"ID":"1556422b-130c-4bf1-8010-ee7c6ea5f662","Type":"ContainerDied","Data":"74490d463f719b89303f16e0bf99043fe1efd7d28d1b6c3e46248505c6a1358f"} Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.034992 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74490d463f719b89303f16e0bf99043fe1efd7d28d1b6c3e46248505c6a1358f" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.035073 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6wv8k" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.048964 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjvxm\" (UniqueName: \"kubernetes.io/projected/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-kube-api-access-wjvxm\") pod \"mysqld-exporter-openstack-cell1-db-create-rx69z\" (UID: \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.050870 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-rx69z\" (UID: \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.152508 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjvxm\" (UniqueName: \"kubernetes.io/projected/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-kube-api-access-wjvxm\") pod \"mysqld-exporter-openstack-cell1-db-create-rx69z\" (UID: \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.152609 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-rx69z\" (UID: \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.154186 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-rx69z\" (UID: \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.177502 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjvxm\" (UniqueName: \"kubernetes.io/projected/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-kube-api-access-wjvxm\") pod \"mysqld-exporter-openstack-cell1-db-create-rx69z\" (UID: \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.193131 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-8f42-account-create-update-2mnqm"] Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.194650 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.198762 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.213125 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-8f42-account-create-update-2mnqm"] Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.264417 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efd8e63-52ce-4ed9-baa0-c1a34b684766-operator-scripts\") pod \"mysqld-exporter-8f42-account-create-update-2mnqm\" (UID: \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\") " pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.264516 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbp96\" (UniqueName: \"kubernetes.io/projected/2efd8e63-52ce-4ed9-baa0-c1a34b684766-kube-api-access-hbp96\") pod \"mysqld-exporter-8f42-account-create-update-2mnqm\" (UID: \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\") " pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.305402 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.368144 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efd8e63-52ce-4ed9-baa0-c1a34b684766-operator-scripts\") pod \"mysqld-exporter-8f42-account-create-update-2mnqm\" (UID: \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\") " pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.371490 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbp96\" (UniqueName: \"kubernetes.io/projected/2efd8e63-52ce-4ed9-baa0-c1a34b684766-kube-api-access-hbp96\") pod \"mysqld-exporter-8f42-account-create-update-2mnqm\" (UID: \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\") " pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.369453 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efd8e63-52ce-4ed9-baa0-c1a34b684766-operator-scripts\") pod \"mysqld-exporter-8f42-account-create-update-2mnqm\" (UID: \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\") " pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.393160 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbp96\" (UniqueName: \"kubernetes.io/projected/2efd8e63-52ce-4ed9-baa0-c1a34b684766-kube-api-access-hbp96\") pod \"mysqld-exporter-8f42-account-create-update-2mnqm\" (UID: \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\") " pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.574981 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.881387 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gbbh7-config-q8574"] Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.912899 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gbbh7-config-q8574"] Dec 03 00:47:43 crc kubenswrapper[4912]: I1203 00:47:43.940392 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-rx69z"] Dec 03 00:47:44 crc kubenswrapper[4912]: I1203 00:47:44.053356 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" event={"ID":"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7","Type":"ContainerStarted","Data":"81f1e27c96397d57d668807cc61d6d9af66dcc751a0f068d9955b4e82918ad6d"} Dec 03 00:47:44 crc kubenswrapper[4912]: I1203 00:47:44.389180 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-8f42-account-create-update-2mnqm"] Dec 03 00:47:44 crc kubenswrapper[4912]: I1203 00:47:44.479600 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 00:47:44 crc kubenswrapper[4912]: I1203 00:47:44.589695 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="112fcd64-ec5f-492c-94ee-731a66de6b63" path="/var/lib/kubelet/pods/112fcd64-ec5f-492c-94ee-731a66de6b63/volumes" Dec 03 00:47:45 crc kubenswrapper[4912]: I1203 00:47:45.004300 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 00:47:45 crc kubenswrapper[4912]: I1203 00:47:45.068246 4912 generic.go:334] "Generic (PLEG): container finished" podID="f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7" containerID="bc63152b5763f5c7b2480ea95d1884cf7987930c728804f066aea3f8e55ec7df" exitCode=0 Dec 03 00:47:45 crc kubenswrapper[4912]: I1203 00:47:45.068367 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" event={"ID":"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7","Type":"ContainerDied","Data":"bc63152b5763f5c7b2480ea95d1884cf7987930c728804f066aea3f8e55ec7df"} Dec 03 00:47:45 crc kubenswrapper[4912]: I1203 00:47:45.070332 4912 generic.go:334] "Generic (PLEG): container finished" podID="2efd8e63-52ce-4ed9-baa0-c1a34b684766" containerID="5b05c942a3905fb2a98c88eaaa3619c19235d97157d3c64423680cfe622581bf" exitCode=0 Dec 03 00:47:45 crc kubenswrapper[4912]: I1203 00:47:45.070390 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" event={"ID":"2efd8e63-52ce-4ed9-baa0-c1a34b684766","Type":"ContainerDied","Data":"5b05c942a3905fb2a98c88eaaa3619c19235d97157d3c64423680cfe622581bf"} Dec 03 00:47:45 crc kubenswrapper[4912]: I1203 00:47:45.070424 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" event={"ID":"2efd8e63-52ce-4ed9-baa0-c1a34b684766","Type":"ContainerStarted","Data":"2916de18a2f5a2e718dd239e9a5d4c1e67cd36ffcef37136427c870d237206e3"} Dec 03 00:47:45 crc kubenswrapper[4912]: I1203 00:47:45.963599 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:45 crc kubenswrapper[4912]: I1203 00:47:45.995473 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/761ebfbd-144d-42cc-ac4f-09a0c49ec7bb-etc-swift\") pod \"swift-storage-0\" (UID: \"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb\") " pod="openstack/swift-storage-0" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.046899 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.614962 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.619937 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.691756 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjvxm\" (UniqueName: \"kubernetes.io/projected/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-kube-api-access-wjvxm\") pod \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\" (UID: \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\") " Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.691961 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-operator-scripts\") pod \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\" (UID: \"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7\") " Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.692094 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbp96\" (UniqueName: \"kubernetes.io/projected/2efd8e63-52ce-4ed9-baa0-c1a34b684766-kube-api-access-hbp96\") pod \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\" (UID: \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\") " Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.692331 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efd8e63-52ce-4ed9-baa0-c1a34b684766-operator-scripts\") pod \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\" (UID: \"2efd8e63-52ce-4ed9-baa0-c1a34b684766\") " Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.698254 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2efd8e63-52ce-4ed9-baa0-c1a34b684766-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2efd8e63-52ce-4ed9-baa0-c1a34b684766" (UID: "2efd8e63-52ce-4ed9-baa0-c1a34b684766"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.698488 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7" (UID: "f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.725736 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2efd8e63-52ce-4ed9-baa0-c1a34b684766-kube-api-access-hbp96" (OuterVolumeSpecName: "kube-api-access-hbp96") pod "2efd8e63-52ce-4ed9-baa0-c1a34b684766" (UID: "2efd8e63-52ce-4ed9-baa0-c1a34b684766"). InnerVolumeSpecName "kube-api-access-hbp96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.725869 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-kube-api-access-wjvxm" (OuterVolumeSpecName: "kube-api-access-wjvxm") pod "f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7" (UID: "f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7"). InnerVolumeSpecName "kube-api-access-wjvxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.795081 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbp96\" (UniqueName: \"kubernetes.io/projected/2efd8e63-52ce-4ed9-baa0-c1a34b684766-kube-api-access-hbp96\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.795122 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2efd8e63-52ce-4ed9-baa0-c1a34b684766-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.795132 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjvxm\" (UniqueName: \"kubernetes.io/projected/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-kube-api-access-wjvxm\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.795141 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:46 crc kubenswrapper[4912]: I1203 00:47:46.942274 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 00:47:46 crc kubenswrapper[4912]: W1203 00:47:46.950247 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod761ebfbd_144d_42cc_ac4f_09a0c49ec7bb.slice/crio-9774646169fb029f90f5fecd8407adfa9338291d825837c0717b4cbae9907533 WatchSource:0}: Error finding container 9774646169fb029f90f5fecd8407adfa9338291d825837c0717b4cbae9907533: Status 404 returned error can't find the container with id 9774646169fb029f90f5fecd8407adfa9338291d825837c0717b4cbae9907533 Dec 03 00:47:47 crc kubenswrapper[4912]: I1203 00:47:47.091823 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"9774646169fb029f90f5fecd8407adfa9338291d825837c0717b4cbae9907533"} Dec 03 00:47:47 crc kubenswrapper[4912]: I1203 00:47:47.094038 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" event={"ID":"f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7","Type":"ContainerDied","Data":"81f1e27c96397d57d668807cc61d6d9af66dcc751a0f068d9955b4e82918ad6d"} Dec 03 00:47:47 crc kubenswrapper[4912]: I1203 00:47:47.094101 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81f1e27c96397d57d668807cc61d6d9af66dcc751a0f068d9955b4e82918ad6d" Dec 03 00:47:47 crc kubenswrapper[4912]: I1203 00:47:47.094189 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-rx69z" Dec 03 00:47:47 crc kubenswrapper[4912]: I1203 00:47:47.102263 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" event={"ID":"2efd8e63-52ce-4ed9-baa0-c1a34b684766","Type":"ContainerDied","Data":"2916de18a2f5a2e718dd239e9a5d4c1e67cd36ffcef37136427c870d237206e3"} Dec 03 00:47:47 crc kubenswrapper[4912]: I1203 00:47:47.102309 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-8f42-account-create-update-2mnqm" Dec 03 00:47:47 crc kubenswrapper[4912]: I1203 00:47:47.102333 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2916de18a2f5a2e718dd239e9a5d4c1e67cd36ffcef37136427c870d237206e3" Dec 03 00:47:47 crc kubenswrapper[4912]: I1203 00:47:47.845724 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.010712 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.354489 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-shcjq"] Dec 03 00:47:48 crc kubenswrapper[4912]: E1203 00:47:48.355284 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2efd8e63-52ce-4ed9-baa0-c1a34b684766" containerName="mariadb-account-create-update" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.355299 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2efd8e63-52ce-4ed9-baa0-c1a34b684766" containerName="mariadb-account-create-update" Dec 03 00:47:48 crc kubenswrapper[4912]: E1203 00:47:48.355317 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7" containerName="mariadb-database-create" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.355323 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7" containerName="mariadb-database-create" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.355511 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7" containerName="mariadb-database-create" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.355528 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2efd8e63-52ce-4ed9-baa0-c1a34b684766" containerName="mariadb-account-create-update" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.356879 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-shcjq" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.385316 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-shcjq"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.443305 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p62g\" (UniqueName: \"kubernetes.io/projected/53898bac-8281-4e12-8caf-c8149ed14a4c-kube-api-access-9p62g\") pod \"cinder-db-create-shcjq\" (UID: \"53898bac-8281-4e12-8caf-c8149ed14a4c\") " pod="openstack/cinder-db-create-shcjq" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.443382 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53898bac-8281-4e12-8caf-c8149ed14a4c-operator-scripts\") pod \"cinder-db-create-shcjq\" (UID: \"53898bac-8281-4e12-8caf-c8149ed14a4c\") " pod="openstack/cinder-db-create-shcjq" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.452242 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-rbqgg"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.456306 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rbqgg" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.475873 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rbqgg"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.548200 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p62g\" (UniqueName: \"kubernetes.io/projected/53898bac-8281-4e12-8caf-c8149ed14a4c-kube-api-access-9p62g\") pod \"cinder-db-create-shcjq\" (UID: \"53898bac-8281-4e12-8caf-c8149ed14a4c\") " pod="openstack/cinder-db-create-shcjq" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.548284 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk798\" (UniqueName: \"kubernetes.io/projected/1edf7b4a-adf2-4e25-ac63-29900c38445a-kube-api-access-vk798\") pod \"barbican-db-create-rbqgg\" (UID: \"1edf7b4a-adf2-4e25-ac63-29900c38445a\") " pod="openstack/barbican-db-create-rbqgg" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.548329 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf7b4a-adf2-4e25-ac63-29900c38445a-operator-scripts\") pod \"barbican-db-create-rbqgg\" (UID: \"1edf7b4a-adf2-4e25-ac63-29900c38445a\") " pod="openstack/barbican-db-create-rbqgg" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.548367 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53898bac-8281-4e12-8caf-c8149ed14a4c-operator-scripts\") pod \"cinder-db-create-shcjq\" (UID: \"53898bac-8281-4e12-8caf-c8149ed14a4c\") " pod="openstack/cinder-db-create-shcjq" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.552924 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53898bac-8281-4e12-8caf-c8149ed14a4c-operator-scripts\") pod \"cinder-db-create-shcjq\" (UID: \"53898bac-8281-4e12-8caf-c8149ed14a4c\") " pod="openstack/cinder-db-create-shcjq" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.593469 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p62g\" (UniqueName: \"kubernetes.io/projected/53898bac-8281-4e12-8caf-c8149ed14a4c-kube-api-access-9p62g\") pod \"cinder-db-create-shcjq\" (UID: \"53898bac-8281-4e12-8caf-c8149ed14a4c\") " pod="openstack/cinder-db-create-shcjq" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.623894 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-prdcm"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.628323 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-prdcm" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.642118 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.644210 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.651577 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk798\" (UniqueName: \"kubernetes.io/projected/1edf7b4a-adf2-4e25-ac63-29900c38445a-kube-api-access-vk798\") pod \"barbican-db-create-rbqgg\" (UID: \"1edf7b4a-adf2-4e25-ac63-29900c38445a\") " pod="openstack/barbican-db-create-rbqgg" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.651656 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf7b4a-adf2-4e25-ac63-29900c38445a-operator-scripts\") pod \"barbican-db-create-rbqgg\" (UID: \"1edf7b4a-adf2-4e25-ac63-29900c38445a\") " pod="openstack/barbican-db-create-rbqgg" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.651729 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00bd8d98-9354-44b8-a44d-76027a084f3c-operator-scripts\") pod \"heat-db-create-prdcm\" (UID: \"00bd8d98-9354-44b8-a44d-76027a084f3c\") " pod="openstack/heat-db-create-prdcm" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.651776 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7q26\" (UniqueName: \"kubernetes.io/projected/00bd8d98-9354-44b8-a44d-76027a084f3c-kube-api-access-s7q26\") pod \"heat-db-create-prdcm\" (UID: \"00bd8d98-9354-44b8-a44d-76027a084f3c\") " pod="openstack/heat-db-create-prdcm" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.653011 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf7b4a-adf2-4e25-ac63-29900c38445a-operator-scripts\") pod \"barbican-db-create-rbqgg\" (UID: \"1edf7b4a-adf2-4e25-ac63-29900c38445a\") " pod="openstack/barbican-db-create-rbqgg" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.655645 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.656979 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-prdcm"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.686547 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-shcjq" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.696526 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.722397 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk798\" (UniqueName: \"kubernetes.io/projected/1edf7b4a-adf2-4e25-ac63-29900c38445a-kube-api-access-vk798\") pod \"barbican-db-create-rbqgg\" (UID: \"1edf7b4a-adf2-4e25-ac63-29900c38445a\") " pod="openstack/barbican-db-create-rbqgg" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.749128 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-j97dl"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.750929 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.754215 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cgq9\" (UniqueName: \"kubernetes.io/projected/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-kube-api-access-8cgq9\") pod \"mysqld-exporter-0\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.754331 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-config-data\") pod \"mysqld-exporter-0\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.754385 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00bd8d98-9354-44b8-a44d-76027a084f3c-operator-scripts\") pod \"heat-db-create-prdcm\" (UID: \"00bd8d98-9354-44b8-a44d-76027a084f3c\") " pod="openstack/heat-db-create-prdcm" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.754445 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7q26\" (UniqueName: \"kubernetes.io/projected/00bd8d98-9354-44b8-a44d-76027a084f3c-kube-api-access-s7q26\") pod \"heat-db-create-prdcm\" (UID: \"00bd8d98-9354-44b8-a44d-76027a084f3c\") " pod="openstack/heat-db-create-prdcm" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.754495 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.755387 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00bd8d98-9354-44b8-a44d-76027a084f3c-operator-scripts\") pod \"heat-db-create-prdcm\" (UID: \"00bd8d98-9354-44b8-a44d-76027a084f3c\") " pod="openstack/heat-db-create-prdcm" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.764356 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.764608 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.764814 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-f2v25" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.765069 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.781940 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rbqgg" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.796859 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-j97dl"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.800486 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7q26\" (UniqueName: \"kubernetes.io/projected/00bd8d98-9354-44b8-a44d-76027a084f3c-kube-api-access-s7q26\") pod \"heat-db-create-prdcm\" (UID: \"00bd8d98-9354-44b8-a44d-76027a084f3c\") " pod="openstack/heat-db-create-prdcm" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.833622 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-852d-account-create-update-l7bkl"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.836087 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.839757 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.853217 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-852d-account-create-update-l7bkl"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.857174 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-combined-ca-bundle\") pod \"keystone-db-sync-j97dl\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.857241 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-config-data\") pod \"keystone-db-sync-j97dl\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.857313 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cgq9\" (UniqueName: \"kubernetes.io/projected/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-kube-api-access-8cgq9\") pod \"mysqld-exporter-0\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.857352 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnfkn\" (UniqueName: \"kubernetes.io/projected/6013018a-f205-4218-a8b2-42ea78b05aaa-kube-api-access-dnfkn\") pod \"heat-852d-account-create-update-l7bkl\" (UID: \"6013018a-f205-4218-a8b2-42ea78b05aaa\") " pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.859814 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-config-data\") pod \"mysqld-exporter-0\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.859878 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5tgb\" (UniqueName: \"kubernetes.io/projected/039cc29c-336b-4c17-9d7d-f08488d07478-kube-api-access-k5tgb\") pod \"keystone-db-sync-j97dl\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.860191 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6013018a-f205-4218-a8b2-42ea78b05aaa-operator-scripts\") pod \"heat-852d-account-create-update-l7bkl\" (UID: \"6013018a-f205-4218-a8b2-42ea78b05aaa\") " pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.866183 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-config-data\") pod \"mysqld-exporter-0\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.860275 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.871526 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.881608 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-4039-account-create-update-g9m7m"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.885016 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.889950 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.891089 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cgq9\" (UniqueName: \"kubernetes.io/projected/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-kube-api-access-8cgq9\") pod \"mysqld-exporter-0\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " pod="openstack/mysqld-exporter-0" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.905610 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4039-account-create-update-g9m7m"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.970020 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-prdcm" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.973059 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpdt4\" (UniqueName: \"kubernetes.io/projected/30895cbf-b323-4348-962b-32e976dbed88-kube-api-access-rpdt4\") pod \"cinder-4039-account-create-update-g9m7m\" (UID: \"30895cbf-b323-4348-962b-32e976dbed88\") " pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.973157 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5tgb\" (UniqueName: \"kubernetes.io/projected/039cc29c-336b-4c17-9d7d-f08488d07478-kube-api-access-k5tgb\") pod \"keystone-db-sync-j97dl\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.973277 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6013018a-f205-4218-a8b2-42ea78b05aaa-operator-scripts\") pod \"heat-852d-account-create-update-l7bkl\" (UID: \"6013018a-f205-4218-a8b2-42ea78b05aaa\") " pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.973334 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-combined-ca-bundle\") pod \"keystone-db-sync-j97dl\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.973362 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-config-data\") pod \"keystone-db-sync-j97dl\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.973403 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnfkn\" (UniqueName: \"kubernetes.io/projected/6013018a-f205-4218-a8b2-42ea78b05aaa-kube-api-access-dnfkn\") pod \"heat-852d-account-create-update-l7bkl\" (UID: \"6013018a-f205-4218-a8b2-42ea78b05aaa\") " pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.973466 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30895cbf-b323-4348-962b-32e976dbed88-operator-scripts\") pod \"cinder-4039-account-create-update-g9m7m\" (UID: \"30895cbf-b323-4348-962b-32e976dbed88\") " pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.975403 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6013018a-f205-4218-a8b2-42ea78b05aaa-operator-scripts\") pod \"heat-852d-account-create-update-l7bkl\" (UID: \"6013018a-f205-4218-a8b2-42ea78b05aaa\") " pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.985562 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-htck5"] Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.986484 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-combined-ca-bundle\") pod \"keystone-db-sync-j97dl\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.987879 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-htck5" Dec 03 00:47:48 crc kubenswrapper[4912]: I1203 00:47:48.999063 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-config-data\") pod \"keystone-db-sync-j97dl\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.007869 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.015599 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5tgb\" (UniqueName: \"kubernetes.io/projected/039cc29c-336b-4c17-9d7d-f08488d07478-kube-api-access-k5tgb\") pod \"keystone-db-sync-j97dl\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.040371 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnfkn\" (UniqueName: \"kubernetes.io/projected/6013018a-f205-4218-a8b2-42ea78b05aaa-kube-api-access-dnfkn\") pod \"heat-852d-account-create-update-l7bkl\" (UID: \"6013018a-f205-4218-a8b2-42ea78b05aaa\") " pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.062091 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-htck5"] Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.076012 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpdt4\" (UniqueName: \"kubernetes.io/projected/30895cbf-b323-4348-962b-32e976dbed88-kube-api-access-rpdt4\") pod \"cinder-4039-account-create-update-g9m7m\" (UID: \"30895cbf-b323-4348-962b-32e976dbed88\") " pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.076151 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gpfl\" (UniqueName: \"kubernetes.io/projected/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-kube-api-access-5gpfl\") pod \"neutron-db-create-htck5\" (UID: \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\") " pod="openstack/neutron-db-create-htck5" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.076226 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-operator-scripts\") pod \"neutron-db-create-htck5\" (UID: \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\") " pod="openstack/neutron-db-create-htck5" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.076269 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30895cbf-b323-4348-962b-32e976dbed88-operator-scripts\") pod \"cinder-4039-account-create-update-g9m7m\" (UID: \"30895cbf-b323-4348-962b-32e976dbed88\") " pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.077371 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30895cbf-b323-4348-962b-32e976dbed88-operator-scripts\") pod \"cinder-4039-account-create-update-g9m7m\" (UID: \"30895cbf-b323-4348-962b-32e976dbed88\") " pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.077444 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-65a2-account-create-update-52ck4"] Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.079028 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.086132 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-65a2-account-create-update-52ck4"] Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.116473 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.127641 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-j97dl" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.128318 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpdt4\" (UniqueName: \"kubernetes.io/projected/30895cbf-b323-4348-962b-32e976dbed88-kube-api-access-rpdt4\") pod \"cinder-4039-account-create-update-g9m7m\" (UID: \"30895cbf-b323-4348-962b-32e976dbed88\") " pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.161910 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.179019 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gpfl\" (UniqueName: \"kubernetes.io/projected/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-kube-api-access-5gpfl\") pod \"neutron-db-create-htck5\" (UID: \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\") " pod="openstack/neutron-db-create-htck5" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.179119 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-operator-scripts\") pod \"neutron-db-create-htck5\" (UID: \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\") " pod="openstack/neutron-db-create-htck5" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.179232 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7594e6bc-7dc5-44cc-8810-c1e972e71764-operator-scripts\") pod \"barbican-65a2-account-create-update-52ck4\" (UID: \"7594e6bc-7dc5-44cc-8810-c1e972e71764\") " pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.179268 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snxmt\" (UniqueName: \"kubernetes.io/projected/7594e6bc-7dc5-44cc-8810-c1e972e71764-kube-api-access-snxmt\") pod \"barbican-65a2-account-create-update-52ck4\" (UID: \"7594e6bc-7dc5-44cc-8810-c1e972e71764\") " pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.187696 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-operator-scripts\") pod \"neutron-db-create-htck5\" (UID: \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\") " pod="openstack/neutron-db-create-htck5" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.216224 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gpfl\" (UniqueName: \"kubernetes.io/projected/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-kube-api-access-5gpfl\") pod \"neutron-db-create-htck5\" (UID: \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\") " pod="openstack/neutron-db-create-htck5" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.236166 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5df8-account-create-update-jblhz"] Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.237774 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.243999 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.275945 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.281186 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7594e6bc-7dc5-44cc-8810-c1e972e71764-operator-scripts\") pod \"barbican-65a2-account-create-update-52ck4\" (UID: \"7594e6bc-7dc5-44cc-8810-c1e972e71764\") " pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.281229 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snxmt\" (UniqueName: \"kubernetes.io/projected/7594e6bc-7dc5-44cc-8810-c1e972e71764-kube-api-access-snxmt\") pod \"barbican-65a2-account-create-update-52ck4\" (UID: \"7594e6bc-7dc5-44cc-8810-c1e972e71764\") " pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.281307 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2lm4\" (UniqueName: \"kubernetes.io/projected/2167d1d7-c169-47a1-ab99-6de7765dc790-kube-api-access-l2lm4\") pod \"neutron-5df8-account-create-update-jblhz\" (UID: \"2167d1d7-c169-47a1-ab99-6de7765dc790\") " pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.281363 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2167d1d7-c169-47a1-ab99-6de7765dc790-operator-scripts\") pod \"neutron-5df8-account-create-update-jblhz\" (UID: \"2167d1d7-c169-47a1-ab99-6de7765dc790\") " pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.282039 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7594e6bc-7dc5-44cc-8810-c1e972e71764-operator-scripts\") pod \"barbican-65a2-account-create-update-52ck4\" (UID: \"7594e6bc-7dc5-44cc-8810-c1e972e71764\") " pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.282301 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5df8-account-create-update-jblhz"] Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.321855 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snxmt\" (UniqueName: \"kubernetes.io/projected/7594e6bc-7dc5-44cc-8810-c1e972e71764-kube-api-access-snxmt\") pod \"barbican-65a2-account-create-update-52ck4\" (UID: \"7594e6bc-7dc5-44cc-8810-c1e972e71764\") " pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.343040 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-htck5" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.382933 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2lm4\" (UniqueName: \"kubernetes.io/projected/2167d1d7-c169-47a1-ab99-6de7765dc790-kube-api-access-l2lm4\") pod \"neutron-5df8-account-create-update-jblhz\" (UID: \"2167d1d7-c169-47a1-ab99-6de7765dc790\") " pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.383003 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2167d1d7-c169-47a1-ab99-6de7765dc790-operator-scripts\") pod \"neutron-5df8-account-create-update-jblhz\" (UID: \"2167d1d7-c169-47a1-ab99-6de7765dc790\") " pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.383840 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2167d1d7-c169-47a1-ab99-6de7765dc790-operator-scripts\") pod \"neutron-5df8-account-create-update-jblhz\" (UID: \"2167d1d7-c169-47a1-ab99-6de7765dc790\") " pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.412788 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.415159 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2lm4\" (UniqueName: \"kubernetes.io/projected/2167d1d7-c169-47a1-ab99-6de7765dc790-kube-api-access-l2lm4\") pod \"neutron-5df8-account-create-update-jblhz\" (UID: \"2167d1d7-c169-47a1-ab99-6de7765dc790\") " pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:47:49 crc kubenswrapper[4912]: I1203 00:47:49.574760 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.246249 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"b29fe5b802586035f404c23b382d24a40cb500417dd58ea9f66609a0cee13c6b"} Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.246922 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"01c4585324afdff8f1d46aaaba5f24c11e2f5a04f52fbc99945d18ad60f73f98"} Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.246934 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"6830ea6b46ed02a6efe125ee2f5af085fd43fd9177604769bbbad1bb6645e634"} Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.314525 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-shcjq"] Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.321230 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-prdcm"] Dec 03 00:47:51 crc kubenswrapper[4912]: W1203 00:47:51.323071 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53898bac_8281_4e12_8caf_c8149ed14a4c.slice/crio-45bc1cb880f94a9d4196d6cfda8d0f1ae6fee863ddaea87a11ab3acaefb31d46 WatchSource:0}: Error finding container 45bc1cb880f94a9d4196d6cfda8d0f1ae6fee863ddaea87a11ab3acaefb31d46: Status 404 returned error can't find the container with id 45bc1cb880f94a9d4196d6cfda8d0f1ae6fee863ddaea87a11ab3acaefb31d46 Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.352462 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-rbqgg"] Dec 03 00:47:51 crc kubenswrapper[4912]: W1203 00:47:51.373311 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00bd8d98_9354_44b8_a44d_76027a084f3c.slice/crio-e6540252e31fcd656287119b60ee34183f3e03810a30c940344fa2a906776dec WatchSource:0}: Error finding container e6540252e31fcd656287119b60ee34183f3e03810a30c940344fa2a906776dec: Status 404 returned error can't find the container with id e6540252e31fcd656287119b60ee34183f3e03810a30c940344fa2a906776dec Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.383576 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-j97dl"] Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.410929 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5df8-account-create-update-jblhz"] Dec 03 00:47:51 crc kubenswrapper[4912]: W1203 00:47:51.457250 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2167d1d7_c169_47a1_ab99_6de7765dc790.slice/crio-701feca568fe83c9e2ddb19b9bb10ddf282f4c678fa9eb5b932e5d58050a42f1 WatchSource:0}: Error finding container 701feca568fe83c9e2ddb19b9bb10ddf282f4c678fa9eb5b932e5d58050a42f1: Status 404 returned error can't find the container with id 701feca568fe83c9e2ddb19b9bb10ddf282f4c678fa9eb5b932e5d58050a42f1 Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.715478 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.743541 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-htck5"] Dec 03 00:47:51 crc kubenswrapper[4912]: W1203 00:47:51.746255 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6013018a_f205_4218_a8b2_42ea78b05aaa.slice/crio-102bcb6fd20dce5b8decb7de3d54773d464de843910d9e850aa4277848148f27 WatchSource:0}: Error finding container 102bcb6fd20dce5b8decb7de3d54773d464de843910d9e850aa4277848148f27: Status 404 returned error can't find the container with id 102bcb6fd20dce5b8decb7de3d54773d464de843910d9e850aa4277848148f27 Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.758250 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-852d-account-create-update-l7bkl"] Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.775672 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-65a2-account-create-update-52ck4"] Dec 03 00:47:51 crc kubenswrapper[4912]: I1203 00:47:51.823269 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-4039-account-create-update-g9m7m"] Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.273389 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-htck5" event={"ID":"5c0f80dc-12ee-4a46-bf09-eceed25e25d1","Type":"ContainerStarted","Data":"4827ee823b9b2839aa8524f743baddb708c3a0aa55ee21b1d4b7563101db442f"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.277126 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"0b261d5855ab5e569dc389b0d74f9ea3a68f9581f95539f19ae2fbf2fbac58aa"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.279307 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1","Type":"ContainerStarted","Data":"db2ded104971550f60bc9edcf0fdc1aedb6c4f95bfeca43fe8e5119e4d1d0c12"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.281172 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-j97dl" event={"ID":"039cc29c-336b-4c17-9d7d-f08488d07478","Type":"ContainerStarted","Data":"51d01a2ae22ce9165f09c225386c0da88c7898f1ed21d7a908609446983d4023"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.290108 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df8-account-create-update-jblhz" event={"ID":"2167d1d7-c169-47a1-ab99-6de7765dc790","Type":"ContainerStarted","Data":"701feca568fe83c9e2ddb19b9bb10ddf282f4c678fa9eb5b932e5d58050a42f1"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.297248 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-shcjq" event={"ID":"53898bac-8281-4e12-8caf-c8149ed14a4c","Type":"ContainerStarted","Data":"3048d2a3a377cdd828e0687fec238d6794925143e3639fb9306dbb3e8ef6cdde"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.297325 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-shcjq" event={"ID":"53898bac-8281-4e12-8caf-c8149ed14a4c","Type":"ContainerStarted","Data":"45bc1cb880f94a9d4196d6cfda8d0f1ae6fee863ddaea87a11ab3acaefb31d46"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.309630 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4039-account-create-update-g9m7m" event={"ID":"30895cbf-b323-4348-962b-32e976dbed88","Type":"ContainerStarted","Data":"a09290248cb820899e91bcfdc8a5f6010c2031c8332210d4063d1408f46047fb"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.317089 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-852d-account-create-update-l7bkl" event={"ID":"6013018a-f205-4218-a8b2-42ea78b05aaa","Type":"ContainerStarted","Data":"102bcb6fd20dce5b8decb7de3d54773d464de843910d9e850aa4277848148f27"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.336813 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rbqgg" event={"ID":"1edf7b4a-adf2-4e25-ac63-29900c38445a","Type":"ContainerStarted","Data":"9f623134169317a8bf47fcb248283e4afcde2f88f3e6f1d9a232b152158ff021"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.336866 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rbqgg" event={"ID":"1edf7b4a-adf2-4e25-ac63-29900c38445a","Type":"ContainerStarted","Data":"335cc536bef752681a2e145c9e55b599334d74bd394a178fdd5641a603a814ce"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.364872 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-prdcm" event={"ID":"00bd8d98-9354-44b8-a44d-76027a084f3c","Type":"ContainerStarted","Data":"e6540252e31fcd656287119b60ee34183f3e03810a30c940344fa2a906776dec"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.373652 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-65a2-account-create-update-52ck4" event={"ID":"7594e6bc-7dc5-44cc-8810-c1e972e71764","Type":"ContainerStarted","Data":"ff2bb21ac95621138631ce10968d770ec9cd91f2ceac36988585f6cc07398131"} Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.388854 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-shcjq" podStartSLOduration=4.3888269619999996 podStartE2EDuration="4.388826962s" podCreationTimestamp="2025-12-03 00:47:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:52.328573253 +0000 UTC m=+1457.970593823" watchObservedRunningTime="2025-12-03 00:47:52.388826962 +0000 UTC m=+1458.030847522" Dec 03 00:47:52 crc kubenswrapper[4912]: I1203 00:47:52.395557 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-rbqgg" podStartSLOduration=4.395519701 podStartE2EDuration="4.395519701s" podCreationTimestamp="2025-12-03 00:47:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:47:52.386633374 +0000 UTC m=+1458.028653924" watchObservedRunningTime="2025-12-03 00:47:52.395519701 +0000 UTC m=+1458.037540291" Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.424528 4912 generic.go:334] "Generic (PLEG): container finished" podID="53898bac-8281-4e12-8caf-c8149ed14a4c" containerID="3048d2a3a377cdd828e0687fec238d6794925143e3639fb9306dbb3e8ef6cdde" exitCode=0 Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.425031 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-shcjq" event={"ID":"53898bac-8281-4e12-8caf-c8149ed14a4c","Type":"ContainerDied","Data":"3048d2a3a377cdd828e0687fec238d6794925143e3639fb9306dbb3e8ef6cdde"} Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.428852 4912 generic.go:334] "Generic (PLEG): container finished" podID="5c0f80dc-12ee-4a46-bf09-eceed25e25d1" containerID="3e2b15f714846231ae7bcf82aafdc8d03d21ac586a603d17914dbabfe32f9cc3" exitCode=0 Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.428932 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-htck5" event={"ID":"5c0f80dc-12ee-4a46-bf09-eceed25e25d1","Type":"ContainerDied","Data":"3e2b15f714846231ae7bcf82aafdc8d03d21ac586a603d17914dbabfe32f9cc3"} Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.434146 4912 generic.go:334] "Generic (PLEG): container finished" podID="30895cbf-b323-4348-962b-32e976dbed88" containerID="a3e22e06b6778bb0eb35826c62cecb496b0bac44d109486edd2c15373a57a2ee" exitCode=0 Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.434260 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4039-account-create-update-g9m7m" event={"ID":"30895cbf-b323-4348-962b-32e976dbed88","Type":"ContainerDied","Data":"a3e22e06b6778bb0eb35826c62cecb496b0bac44d109486edd2c15373a57a2ee"} Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.438857 4912 generic.go:334] "Generic (PLEG): container finished" podID="6013018a-f205-4218-a8b2-42ea78b05aaa" containerID="9f58bc4c1070589d8cd090a43fbacdacda88465c4e21f61cc174d5fafe050899" exitCode=0 Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.438991 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-852d-account-create-update-l7bkl" event={"ID":"6013018a-f205-4218-a8b2-42ea78b05aaa","Type":"ContainerDied","Data":"9f58bc4c1070589d8cd090a43fbacdacda88465c4e21f61cc174d5fafe050899"} Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.441027 4912 generic.go:334] "Generic (PLEG): container finished" podID="2167d1d7-c169-47a1-ab99-6de7765dc790" containerID="dbf85cf9f3adc7e94fdaa3b6eab3f310c481536c24eaa07dd2f23abe0d163ff0" exitCode=0 Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.441100 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df8-account-create-update-jblhz" event={"ID":"2167d1d7-c169-47a1-ab99-6de7765dc790","Type":"ContainerDied","Data":"dbf85cf9f3adc7e94fdaa3b6eab3f310c481536c24eaa07dd2f23abe0d163ff0"} Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.442781 4912 generic.go:334] "Generic (PLEG): container finished" podID="1edf7b4a-adf2-4e25-ac63-29900c38445a" containerID="9f623134169317a8bf47fcb248283e4afcde2f88f3e6f1d9a232b152158ff021" exitCode=0 Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.442874 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rbqgg" event={"ID":"1edf7b4a-adf2-4e25-ac63-29900c38445a","Type":"ContainerDied","Data":"9f623134169317a8bf47fcb248283e4afcde2f88f3e6f1d9a232b152158ff021"} Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.445297 4912 generic.go:334] "Generic (PLEG): container finished" podID="00bd8d98-9354-44b8-a44d-76027a084f3c" containerID="65b597ed4bbc1350c58cf547afdde883dd0adf46b4211e515132b0e0ad6962b8" exitCode=0 Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.445362 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-prdcm" event={"ID":"00bd8d98-9354-44b8-a44d-76027a084f3c","Type":"ContainerDied","Data":"65b597ed4bbc1350c58cf547afdde883dd0adf46b4211e515132b0e0ad6962b8"} Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.452022 4912 generic.go:334] "Generic (PLEG): container finished" podID="7594e6bc-7dc5-44cc-8810-c1e972e71764" containerID="94ab507a74139bebe0993d007af1bdb40614a1c74ee316763ce88acc4c210a9d" exitCode=0 Dec 03 00:47:53 crc kubenswrapper[4912]: I1203 00:47:53.452106 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-65a2-account-create-update-52ck4" event={"ID":"7594e6bc-7dc5-44cc-8810-c1e972e71764","Type":"ContainerDied","Data":"94ab507a74139bebe0993d007af1bdb40614a1c74ee316763ce88acc4c210a9d"} Dec 03 00:47:54 crc kubenswrapper[4912]: I1203 00:47:54.479138 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 00:47:54 crc kubenswrapper[4912]: I1203 00:47:54.485351 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 00:47:55 crc kubenswrapper[4912]: I1203 00:47:55.493868 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 00:47:57 crc kubenswrapper[4912]: I1203 00:47:57.879979 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 00:47:57 crc kubenswrapper[4912]: I1203 00:47:57.880869 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="thanos-sidecar" containerID="cri-o://f99af3ce42b0c9f1459f5bc50c596e75148a7ed8d7be5a36b61b602acbda2801" gracePeriod=600 Dec 03 00:47:57 crc kubenswrapper[4912]: I1203 00:47:57.881001 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="config-reloader" containerID="cri-o://0879a34d96d5a73e15cfbe99d2677a45352f81b80652556076f6cab36d64c53b" gracePeriod=600 Dec 03 00:47:57 crc kubenswrapper[4912]: I1203 00:47:57.881066 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="prometheus" containerID="cri-o://d62cd23fa7798120cbc7fbf3e89799268dfd494b1f4854490741343fa74ec4f8" gracePeriod=600 Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.484793 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.125:9090/-/ready\": dial tcp 10.217.0.125:9090: connect: connection refused" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.540575 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-htck5" event={"ID":"5c0f80dc-12ee-4a46-bf09-eceed25e25d1","Type":"ContainerDied","Data":"4827ee823b9b2839aa8524f743baddb708c3a0aa55ee21b1d4b7563101db442f"} Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.540641 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4827ee823b9b2839aa8524f743baddb708c3a0aa55ee21b1d4b7563101db442f" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.542149 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-4039-account-create-update-g9m7m" event={"ID":"30895cbf-b323-4348-962b-32e976dbed88","Type":"ContainerDied","Data":"a09290248cb820899e91bcfdc8a5f6010c2031c8332210d4063d1408f46047fb"} Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.542210 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a09290248cb820899e91bcfdc8a5f6010c2031c8332210d4063d1408f46047fb" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.543765 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-852d-account-create-update-l7bkl" event={"ID":"6013018a-f205-4218-a8b2-42ea78b05aaa","Type":"ContainerDied","Data":"102bcb6fd20dce5b8decb7de3d54773d464de843910d9e850aa4277848148f27"} Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.543792 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="102bcb6fd20dce5b8decb7de3d54773d464de843910d9e850aa4277848148f27" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.544973 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5df8-account-create-update-jblhz" event={"ID":"2167d1d7-c169-47a1-ab99-6de7765dc790","Type":"ContainerDied","Data":"701feca568fe83c9e2ddb19b9bb10ddf282f4c678fa9eb5b932e5d58050a42f1"} Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.545030 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="701feca568fe83c9e2ddb19b9bb10ddf282f4c678fa9eb5b932e5d58050a42f1" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.547459 4912 generic.go:334] "Generic (PLEG): container finished" podID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerID="f99af3ce42b0c9f1459f5bc50c596e75148a7ed8d7be5a36b61b602acbda2801" exitCode=0 Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.547498 4912 generic.go:334] "Generic (PLEG): container finished" podID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerID="0879a34d96d5a73e15cfbe99d2677a45352f81b80652556076f6cab36d64c53b" exitCode=0 Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.547508 4912 generic.go:334] "Generic (PLEG): container finished" podID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerID="d62cd23fa7798120cbc7fbf3e89799268dfd494b1f4854490741343fa74ec4f8" exitCode=0 Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.547471 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerDied","Data":"f99af3ce42b0c9f1459f5bc50c596e75148a7ed8d7be5a36b61b602acbda2801"} Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.547623 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerDied","Data":"0879a34d96d5a73e15cfbe99d2677a45352f81b80652556076f6cab36d64c53b"} Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.547643 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerDied","Data":"d62cd23fa7798120cbc7fbf3e89799268dfd494b1f4854490741343fa74ec4f8"} Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.552057 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-rbqgg" event={"ID":"1edf7b4a-adf2-4e25-ac63-29900c38445a","Type":"ContainerDied","Data":"335cc536bef752681a2e145c9e55b599334d74bd394a178fdd5641a603a814ce"} Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.552110 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="335cc536bef752681a2e145c9e55b599334d74bd394a178fdd5641a603a814ce" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.554100 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-prdcm" event={"ID":"00bd8d98-9354-44b8-a44d-76027a084f3c","Type":"ContainerDied","Data":"e6540252e31fcd656287119b60ee34183f3e03810a30c940344fa2a906776dec"} Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.554130 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6540252e31fcd656287119b60ee34183f3e03810a30c940344fa2a906776dec" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.766843 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.780491 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.805730 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rbqgg" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.863941 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-htck5" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.887804 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.892763 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vk798\" (UniqueName: \"kubernetes.io/projected/1edf7b4a-adf2-4e25-ac63-29900c38445a-kube-api-access-vk798\") pod \"1edf7b4a-adf2-4e25-ac63-29900c38445a\" (UID: \"1edf7b4a-adf2-4e25-ac63-29900c38445a\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.892875 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnfkn\" (UniqueName: \"kubernetes.io/projected/6013018a-f205-4218-a8b2-42ea78b05aaa-kube-api-access-dnfkn\") pod \"6013018a-f205-4218-a8b2-42ea78b05aaa\" (UID: \"6013018a-f205-4218-a8b2-42ea78b05aaa\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.892930 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf7b4a-adf2-4e25-ac63-29900c38445a-operator-scripts\") pod \"1edf7b4a-adf2-4e25-ac63-29900c38445a\" (UID: \"1edf7b4a-adf2-4e25-ac63-29900c38445a\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.892964 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpdt4\" (UniqueName: \"kubernetes.io/projected/30895cbf-b323-4348-962b-32e976dbed88-kube-api-access-rpdt4\") pod \"30895cbf-b323-4348-962b-32e976dbed88\" (UID: \"30895cbf-b323-4348-962b-32e976dbed88\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.892987 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6013018a-f205-4218-a8b2-42ea78b05aaa-operator-scripts\") pod \"6013018a-f205-4218-a8b2-42ea78b05aaa\" (UID: \"6013018a-f205-4218-a8b2-42ea78b05aaa\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.893008 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30895cbf-b323-4348-962b-32e976dbed88-operator-scripts\") pod \"30895cbf-b323-4348-962b-32e976dbed88\" (UID: \"30895cbf-b323-4348-962b-32e976dbed88\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.895377 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30895cbf-b323-4348-962b-32e976dbed88-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "30895cbf-b323-4348-962b-32e976dbed88" (UID: "30895cbf-b323-4348-962b-32e976dbed88"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.899645 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1edf7b4a-adf2-4e25-ac63-29900c38445a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1edf7b4a-adf2-4e25-ac63-29900c38445a" (UID: "1edf7b4a-adf2-4e25-ac63-29900c38445a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.902013 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-prdcm" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.906039 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1edf7b4a-adf2-4e25-ac63-29900c38445a-kube-api-access-vk798" (OuterVolumeSpecName: "kube-api-access-vk798") pod "1edf7b4a-adf2-4e25-ac63-29900c38445a" (UID: "1edf7b4a-adf2-4e25-ac63-29900c38445a"). InnerVolumeSpecName "kube-api-access-vk798". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.907672 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6013018a-f205-4218-a8b2-42ea78b05aaa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6013018a-f205-4218-a8b2-42ea78b05aaa" (UID: "6013018a-f205-4218-a8b2-42ea78b05aaa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.910988 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6013018a-f205-4218-a8b2-42ea78b05aaa-kube-api-access-dnfkn" (OuterVolumeSpecName: "kube-api-access-dnfkn") pod "6013018a-f205-4218-a8b2-42ea78b05aaa" (UID: "6013018a-f205-4218-a8b2-42ea78b05aaa"). InnerVolumeSpecName "kube-api-access-dnfkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.911107 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30895cbf-b323-4348-962b-32e976dbed88-kube-api-access-rpdt4" (OuterVolumeSpecName: "kube-api-access-rpdt4") pod "30895cbf-b323-4348-962b-32e976dbed88" (UID: "30895cbf-b323-4348-962b-32e976dbed88"). InnerVolumeSpecName "kube-api-access-rpdt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995015 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-operator-scripts\") pod \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\" (UID: \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995119 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7q26\" (UniqueName: \"kubernetes.io/projected/00bd8d98-9354-44b8-a44d-76027a084f3c-kube-api-access-s7q26\") pod \"00bd8d98-9354-44b8-a44d-76027a084f3c\" (UID: \"00bd8d98-9354-44b8-a44d-76027a084f3c\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995161 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2167d1d7-c169-47a1-ab99-6de7765dc790-operator-scripts\") pod \"2167d1d7-c169-47a1-ab99-6de7765dc790\" (UID: \"2167d1d7-c169-47a1-ab99-6de7765dc790\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995192 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2lm4\" (UniqueName: \"kubernetes.io/projected/2167d1d7-c169-47a1-ab99-6de7765dc790-kube-api-access-l2lm4\") pod \"2167d1d7-c169-47a1-ab99-6de7765dc790\" (UID: \"2167d1d7-c169-47a1-ab99-6de7765dc790\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995326 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00bd8d98-9354-44b8-a44d-76027a084f3c-operator-scripts\") pod \"00bd8d98-9354-44b8-a44d-76027a084f3c\" (UID: \"00bd8d98-9354-44b8-a44d-76027a084f3c\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995396 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gpfl\" (UniqueName: \"kubernetes.io/projected/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-kube-api-access-5gpfl\") pod \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\" (UID: \"5c0f80dc-12ee-4a46-bf09-eceed25e25d1\") " Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995923 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vk798\" (UniqueName: \"kubernetes.io/projected/1edf7b4a-adf2-4e25-ac63-29900c38445a-kube-api-access-vk798\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995939 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnfkn\" (UniqueName: \"kubernetes.io/projected/6013018a-f205-4218-a8b2-42ea78b05aaa-kube-api-access-dnfkn\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995948 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1edf7b4a-adf2-4e25-ac63-29900c38445a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995957 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpdt4\" (UniqueName: \"kubernetes.io/projected/30895cbf-b323-4348-962b-32e976dbed88-kube-api-access-rpdt4\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995965 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/30895cbf-b323-4348-962b-32e976dbed88-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.995975 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6013018a-f205-4218-a8b2-42ea78b05aaa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.996600 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:47:59 crc kubenswrapper[4912]: I1203 00:47:59.997011 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c0f80dc-12ee-4a46-bf09-eceed25e25d1" (UID: "5c0f80dc-12ee-4a46-bf09-eceed25e25d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:47:59.997935 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2167d1d7-c169-47a1-ab99-6de7765dc790-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2167d1d7-c169-47a1-ab99-6de7765dc790" (UID: "2167d1d7-c169-47a1-ab99-6de7765dc790"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:47:59.999806 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00bd8d98-9354-44b8-a44d-76027a084f3c-kube-api-access-s7q26" (OuterVolumeSpecName: "kube-api-access-s7q26") pod "00bd8d98-9354-44b8-a44d-76027a084f3c" (UID: "00bd8d98-9354-44b8-a44d-76027a084f3c"). InnerVolumeSpecName "kube-api-access-s7q26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.000494 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-kube-api-access-5gpfl" (OuterVolumeSpecName: "kube-api-access-5gpfl") pod "5c0f80dc-12ee-4a46-bf09-eceed25e25d1" (UID: "5c0f80dc-12ee-4a46-bf09-eceed25e25d1"). InnerVolumeSpecName "kube-api-access-5gpfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.002589 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2167d1d7-c169-47a1-ab99-6de7765dc790-kube-api-access-l2lm4" (OuterVolumeSpecName: "kube-api-access-l2lm4") pod "2167d1d7-c169-47a1-ab99-6de7765dc790" (UID: "2167d1d7-c169-47a1-ab99-6de7765dc790"). InnerVolumeSpecName "kube-api-access-l2lm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.003228 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-shcjq" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.017736 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00bd8d98-9354-44b8-a44d-76027a084f3c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00bd8d98-9354-44b8-a44d-76027a084f3c" (UID: "00bd8d98-9354-44b8-a44d-76027a084f3c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.097834 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p62g\" (UniqueName: \"kubernetes.io/projected/53898bac-8281-4e12-8caf-c8149ed14a4c-kube-api-access-9p62g\") pod \"53898bac-8281-4e12-8caf-c8149ed14a4c\" (UID: \"53898bac-8281-4e12-8caf-c8149ed14a4c\") " Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.097928 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snxmt\" (UniqueName: \"kubernetes.io/projected/7594e6bc-7dc5-44cc-8810-c1e972e71764-kube-api-access-snxmt\") pod \"7594e6bc-7dc5-44cc-8810-c1e972e71764\" (UID: \"7594e6bc-7dc5-44cc-8810-c1e972e71764\") " Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.098080 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53898bac-8281-4e12-8caf-c8149ed14a4c-operator-scripts\") pod \"53898bac-8281-4e12-8caf-c8149ed14a4c\" (UID: \"53898bac-8281-4e12-8caf-c8149ed14a4c\") " Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.098248 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7594e6bc-7dc5-44cc-8810-c1e972e71764-operator-scripts\") pod \"7594e6bc-7dc5-44cc-8810-c1e972e71764\" (UID: \"7594e6bc-7dc5-44cc-8810-c1e972e71764\") " Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.098720 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.098735 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7q26\" (UniqueName: \"kubernetes.io/projected/00bd8d98-9354-44b8-a44d-76027a084f3c-kube-api-access-s7q26\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.098749 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2167d1d7-c169-47a1-ab99-6de7765dc790-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.098758 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2lm4\" (UniqueName: \"kubernetes.io/projected/2167d1d7-c169-47a1-ab99-6de7765dc790-kube-api-access-l2lm4\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.098766 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00bd8d98-9354-44b8-a44d-76027a084f3c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.098776 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gpfl\" (UniqueName: \"kubernetes.io/projected/5c0f80dc-12ee-4a46-bf09-eceed25e25d1-kube-api-access-5gpfl\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.099281 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53898bac-8281-4e12-8caf-c8149ed14a4c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53898bac-8281-4e12-8caf-c8149ed14a4c" (UID: "53898bac-8281-4e12-8caf-c8149ed14a4c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.103143 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53898bac-8281-4e12-8caf-c8149ed14a4c-kube-api-access-9p62g" (OuterVolumeSpecName: "kube-api-access-9p62g") pod "53898bac-8281-4e12-8caf-c8149ed14a4c" (UID: "53898bac-8281-4e12-8caf-c8149ed14a4c"). InnerVolumeSpecName "kube-api-access-9p62g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.103774 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7594e6bc-7dc5-44cc-8810-c1e972e71764-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7594e6bc-7dc5-44cc-8810-c1e972e71764" (UID: "7594e6bc-7dc5-44cc-8810-c1e972e71764"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.104837 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7594e6bc-7dc5-44cc-8810-c1e972e71764-kube-api-access-snxmt" (OuterVolumeSpecName: "kube-api-access-snxmt") pod "7594e6bc-7dc5-44cc-8810-c1e972e71764" (UID: "7594e6bc-7dc5-44cc-8810-c1e972e71764"). InnerVolumeSpecName "kube-api-access-snxmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.200370 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7594e6bc-7dc5-44cc-8810-c1e972e71764-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.200414 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p62g\" (UniqueName: \"kubernetes.io/projected/53898bac-8281-4e12-8caf-c8149ed14a4c-kube-api-access-9p62g\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.200451 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snxmt\" (UniqueName: \"kubernetes.io/projected/7594e6bc-7dc5-44cc-8810-c1e972e71764-kube-api-access-snxmt\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.200464 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53898bac-8281-4e12-8caf-c8149ed14a4c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.612528 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-65a2-account-create-update-52ck4" event={"ID":"7594e6bc-7dc5-44cc-8810-c1e972e71764","Type":"ContainerDied","Data":"ff2bb21ac95621138631ce10968d770ec9cd91f2ceac36988585f6cc07398131"} Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.612933 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff2bb21ac95621138631ce10968d770ec9cd91f2ceac36988585f6cc07398131" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.612771 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-65a2-account-create-update-52ck4" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.616348 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-852d-account-create-update-l7bkl" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.618555 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-prdcm" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.620719 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-4039-account-create-update-g9m7m" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.629578 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-htck5" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.629651 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-shcjq" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.629740 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-shcjq" event={"ID":"53898bac-8281-4e12-8caf-c8149ed14a4c","Type":"ContainerDied","Data":"45bc1cb880f94a9d4196d6cfda8d0f1ae6fee863ddaea87a11ab3acaefb31d46"} Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.629781 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45bc1cb880f94a9d4196d6cfda8d0f1ae6fee863ddaea87a11ab3acaefb31d46" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.629818 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5df8-account-create-update-jblhz" Dec 03 00:48:00 crc kubenswrapper[4912]: I1203 00:48:00.629844 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-rbqgg" Dec 03 00:48:04 crc kubenswrapper[4912]: I1203 00:48:04.479920 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.125:9090/-/ready\": dial tcp 10.217.0.125:9090: connect: connection refused" Dec 03 00:48:08 crc kubenswrapper[4912]: E1203 00:48:08.902332 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 03 00:48:08 crc kubenswrapper[4912]: E1203 00:48:08.905309 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mk7cz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-tnx9g_openstack(e760382f-ba1d-4d51-bdd0-2cd62315c4ae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:48:08 crc kubenswrapper[4912]: E1203 00:48:08.906801 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-tnx9g" podUID="e760382f-ba1d-4d51-bdd0-2cd62315c4ae" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.408511 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.533679 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57brj\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-kube-api-access-57brj\") pod \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.533770 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-thanos-prometheus-http-client-file\") pod \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.533864 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config-out\") pod \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.533886 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-prometheus-metric-storage-rulefiles-0\") pod \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.533937 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-tls-assets\") pod \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.533978 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config\") pod \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.533996 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.534137 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-web-config\") pod \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\" (UID: \"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a\") " Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.534950 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" (UID: "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.542825 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" (UID: "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.544226 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config-out" (OuterVolumeSpecName: "config-out") pod "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" (UID: "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.549789 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" (UID: "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.550580 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config" (OuterVolumeSpecName: "config") pod "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" (UID: "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.553845 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-kube-api-access-57brj" (OuterVolumeSpecName: "kube-api-access-57brj") pod "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" (UID: "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a"). InnerVolumeSpecName "kube-api-access-57brj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.558670 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" (UID: "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.586539 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-web-config" (OuterVolumeSpecName: "web-config") pod "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" (UID: "53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.636688 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57brj\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-kube-api-access-57brj\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.636739 4912 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.636756 4912 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config-out\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.636772 4912 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.636789 4912 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.636801 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.636835 4912 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.636847 4912 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a-web-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.664406 4912 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.740105 4912 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.752115 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-j97dl" event={"ID":"039cc29c-336b-4c17-9d7d-f08488d07478","Type":"ContainerStarted","Data":"73649cd60fde4ac8bc10e0e8b451468eee12d3d2a23ab399e0c2a2461a725bd6"} Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.757561 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"510cb60ad136c7d60a008e18438426b97f2f2717d5a713c3535b7dd37ad1fc40"} Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.757637 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"aa29305081db50f86fdd5756a5468d7642ea04cd1d568850e5203988329cfd28"} Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.757653 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"3c73f7a7fe4a178294f1470a45c8b29ac07599591898444c547f4b1e3dddb357"} Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.760045 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1","Type":"ContainerStarted","Data":"740a3af8062d26930e27ad3a6b7b2a65f5d8cffa7149732694da91302c905cf9"} Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.764568 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a","Type":"ContainerDied","Data":"80dc7755a28ec37661ffcee2c27efedc08f43e0c2c0241515e53bbd765a9dccb"} Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.764661 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.764700 4912 scope.go:117] "RemoveContainer" containerID="f99af3ce42b0c9f1459f5bc50c596e75148a7ed8d7be5a36b61b602acbda2801" Dec 03 00:48:09 crc kubenswrapper[4912]: E1203 00:48:09.766296 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-tnx9g" podUID="e760382f-ba1d-4d51-bdd0-2cd62315c4ae" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.797036 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-j97dl" podStartSLOduration=4.239135844 podStartE2EDuration="21.797014742s" podCreationTimestamp="2025-12-03 00:47:48 +0000 UTC" firstStartedPulling="2025-12-03 00:47:51.432976244 +0000 UTC m=+1457.074996794" lastFinishedPulling="2025-12-03 00:48:08.990855132 +0000 UTC m=+1474.632875692" observedRunningTime="2025-12-03 00:48:09.784659632 +0000 UTC m=+1475.426680192" watchObservedRunningTime="2025-12-03 00:48:09.797014742 +0000 UTC m=+1475.439035302" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.799587 4912 scope.go:117] "RemoveContainer" containerID="0879a34d96d5a73e15cfbe99d2677a45352f81b80652556076f6cab36d64c53b" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.842384 4912 scope.go:117] "RemoveContainer" containerID="d62cd23fa7798120cbc7fbf3e89799268dfd494b1f4854490741343fa74ec4f8" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.844746 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=14.176138576 podStartE2EDuration="21.844732697s" podCreationTimestamp="2025-12-03 00:47:48 +0000 UTC" firstStartedPulling="2025-12-03 00:47:51.796705068 +0000 UTC m=+1457.438725628" lastFinishedPulling="2025-12-03 00:47:59.465299189 +0000 UTC m=+1465.107319749" observedRunningTime="2025-12-03 00:48:09.843073852 +0000 UTC m=+1475.485094402" watchObservedRunningTime="2025-12-03 00:48:09.844732697 +0000 UTC m=+1475.486753257" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.894367 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.924902 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.960639 4912 scope.go:117] "RemoveContainer" containerID="ad44c075b7f7b30fd0825216a0ab3035fce81f0b1a2ab801fe30d418c4b82483" Dec 03 00:48:09 crc kubenswrapper[4912]: I1203 00:48:09.995950 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.015926 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="config-reloader" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.015980 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="config-reloader" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016069 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="prometheus" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016075 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="prometheus" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016090 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30895cbf-b323-4348-962b-32e976dbed88" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016099 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="30895cbf-b323-4348-962b-32e976dbed88" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016119 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c0f80dc-12ee-4a46-bf09-eceed25e25d1" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016126 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c0f80dc-12ee-4a46-bf09-eceed25e25d1" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016141 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6013018a-f205-4218-a8b2-42ea78b05aaa" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016149 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="6013018a-f205-4218-a8b2-42ea78b05aaa" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016172 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2167d1d7-c169-47a1-ab99-6de7765dc790" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016179 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2167d1d7-c169-47a1-ab99-6de7765dc790" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016199 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00bd8d98-9354-44b8-a44d-76027a084f3c" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016207 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="00bd8d98-9354-44b8-a44d-76027a084f3c" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016229 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7594e6bc-7dc5-44cc-8810-c1e972e71764" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016278 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7594e6bc-7dc5-44cc-8810-c1e972e71764" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016321 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="init-config-reloader" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016338 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="init-config-reloader" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016367 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="thanos-sidecar" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016375 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="thanos-sidecar" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016404 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53898bac-8281-4e12-8caf-c8149ed14a4c" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016412 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="53898bac-8281-4e12-8caf-c8149ed14a4c" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: E1203 00:48:10.016523 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1edf7b4a-adf2-4e25-ac63-29900c38445a" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.016538 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1edf7b4a-adf2-4e25-ac63-29900c38445a" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.017391 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="53898bac-8281-4e12-8caf-c8149ed14a4c" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036371 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="thanos-sidecar" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036461 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="6013018a-f205-4218-a8b2-42ea78b05aaa" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036483 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2167d1d7-c169-47a1-ab99-6de7765dc790" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036501 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c0f80dc-12ee-4a46-bf09-eceed25e25d1" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036517 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="30895cbf-b323-4348-962b-32e976dbed88" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036541 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="1edf7b4a-adf2-4e25-ac63-29900c38445a" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036576 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7594e6bc-7dc5-44cc-8810-c1e972e71764" containerName="mariadb-account-create-update" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036591 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="prometheus" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036606 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="00bd8d98-9354-44b8-a44d-76027a084f3c" containerName="mariadb-database-create" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.036625 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" containerName="config-reloader" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.059070 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.059206 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.062770 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.066785 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.066944 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.067013 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.067199 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-w4fkg" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.067928 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.075773 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.168573 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.168635 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.168676 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-config\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.168801 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.168962 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.169179 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.169228 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.169272 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.169346 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.169583 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.169632 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8f67\" (UniqueName: \"kubernetes.io/projected/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-kube-api-access-g8f67\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271517 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271579 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271604 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271642 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271690 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271716 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8f67\" (UniqueName: \"kubernetes.io/projected/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-kube-api-access-g8f67\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271772 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271800 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271835 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-config\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271865 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.271890 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.272016 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.273106 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.278152 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.280343 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.280542 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.284091 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.284356 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.284380 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-config\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.286687 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.288610 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.294708 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8f67\" (UniqueName: \"kubernetes.io/projected/f78bc494-84ea-43ff-b3fc-0af2f10c43b4-kube-api-access-g8f67\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.316874 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"prometheus-metric-storage-0\" (UID: \"f78bc494-84ea-43ff-b3fc-0af2f10c43b4\") " pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.389588 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.600785 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a" path="/var/lib/kubelet/pods/53caa3f1-e58a-4d0f-814d-5aa3cd8dae0a/volumes" Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.785726 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"4b47e0847f728a212bfed3ab8a5ed27d115565e9cd0b759ffc1b338fcc3dc923"} Dec 03 00:48:10 crc kubenswrapper[4912]: I1203 00:48:10.909644 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 00:48:10 crc kubenswrapper[4912]: W1203 00:48:10.922015 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf78bc494_84ea_43ff_b3fc_0af2f10c43b4.slice/crio-27038411e625b67f474e53334e05692620962e44729c6088363d3f94feb9c490 WatchSource:0}: Error finding container 27038411e625b67f474e53334e05692620962e44729c6088363d3f94feb9c490: Status 404 returned error can't find the container with id 27038411e625b67f474e53334e05692620962e44729c6088363d3f94feb9c490 Dec 03 00:48:11 crc kubenswrapper[4912]: I1203 00:48:11.802772 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"1de305a65cf8129ed49c7886de410a3f7b13a1b0d6218467918d79326d3ebff9"} Dec 03 00:48:11 crc kubenswrapper[4912]: I1203 00:48:11.805092 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f78bc494-84ea-43ff-b3fc-0af2f10c43b4","Type":"ContainerStarted","Data":"27038411e625b67f474e53334e05692620962e44729c6088363d3f94feb9c490"} Dec 03 00:48:12 crc kubenswrapper[4912]: I1203 00:48:12.844387 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"08e0d90a13af4da3abee4afb876c91a437746b2ab6427b348d567dd65a67755f"} Dec 03 00:48:12 crc kubenswrapper[4912]: I1203 00:48:12.844781 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"803547ae93af1303885466dc6c713b0cbf1a7a4e2bf06f5c26389ced8e33035e"} Dec 03 00:48:12 crc kubenswrapper[4912]: I1203 00:48:12.844794 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"016a87cf64e32c7d2b2065dcea8f69f1f703369e77824dbd8e17c5ad6334cc9c"} Dec 03 00:48:12 crc kubenswrapper[4912]: I1203 00:48:12.844804 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"62979de349f42cb6386b59abf5425c412d8db9f968a128f7fb1d0d6870b6ede8"} Dec 03 00:48:13 crc kubenswrapper[4912]: I1203 00:48:13.880151 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"a6f9fa110f4507174392e6f7aec359f7eace40aeec54c991b4c9d0fd5ad40fbe"} Dec 03 00:48:13 crc kubenswrapper[4912]: I1203 00:48:13.883736 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"761ebfbd-144d-42cc-ac4f-09a0c49ec7bb","Type":"ContainerStarted","Data":"d47546787fefc2a2a40a287e322b8994c00b8f7a881c038a8de152c47be2d19c"} Dec 03 00:48:13 crc kubenswrapper[4912]: I1203 00:48:13.885943 4912 generic.go:334] "Generic (PLEG): container finished" podID="039cc29c-336b-4c17-9d7d-f08488d07478" containerID="73649cd60fde4ac8bc10e0e8b451468eee12d3d2a23ab399e0c2a2461a725bd6" exitCode=0 Dec 03 00:48:13 crc kubenswrapper[4912]: I1203 00:48:13.886025 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-j97dl" event={"ID":"039cc29c-336b-4c17-9d7d-f08488d07478","Type":"ContainerDied","Data":"73649cd60fde4ac8bc10e0e8b451468eee12d3d2a23ab399e0c2a2461a725bd6"} Dec 03 00:48:13 crc kubenswrapper[4912]: I1203 00:48:13.934381 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.38613017 podStartE2EDuration="1m1.934362707s" podCreationTimestamp="2025-12-03 00:47:12 +0000 UTC" firstStartedPulling="2025-12-03 00:47:46.953168785 +0000 UTC m=+1452.595189345" lastFinishedPulling="2025-12-03 00:48:11.501401322 +0000 UTC m=+1477.143421882" observedRunningTime="2025-12-03 00:48:13.931299225 +0000 UTC m=+1479.573319805" watchObservedRunningTime="2025-12-03 00:48:13.934362707 +0000 UTC m=+1479.576383267" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.254055 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-n2jc6"] Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.256554 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.259151 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.272628 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-n2jc6"] Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.377405 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.377501 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.377610 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-config\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.378031 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.378106 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.378521 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-885df\" (UniqueName: \"kubernetes.io/projected/21bacb6a-16b3-41d4-858c-83e04de8aa9e-kube-api-access-885df\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.480706 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.480782 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.480808 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-config\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.480882 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.480911 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.480999 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-885df\" (UniqueName: \"kubernetes.io/projected/21bacb6a-16b3-41d4-858c-83e04de8aa9e-kube-api-access-885df\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.482224 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.482518 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-config\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.482584 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.482695 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.483156 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.507158 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-885df\" (UniqueName: \"kubernetes.io/projected/21bacb6a-16b3-41d4-858c-83e04de8aa9e-kube-api-access-885df\") pod \"dnsmasq-dns-5c79d794d7-n2jc6\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.598723 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:14 crc kubenswrapper[4912]: I1203 00:48:14.908563 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f78bc494-84ea-43ff-b3fc-0af2f10c43b4","Type":"ContainerStarted","Data":"2a0474e3f923d3125a4b84523cca836e5e5ba4fdfcffc3035b6dbddbbd677e31"} Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.239648 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-n2jc6"] Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.619805 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-j97dl" Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.713411 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-config-data\") pod \"039cc29c-336b-4c17-9d7d-f08488d07478\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.713653 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5tgb\" (UniqueName: \"kubernetes.io/projected/039cc29c-336b-4c17-9d7d-f08488d07478-kube-api-access-k5tgb\") pod \"039cc29c-336b-4c17-9d7d-f08488d07478\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.713732 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-combined-ca-bundle\") pod \"039cc29c-336b-4c17-9d7d-f08488d07478\" (UID: \"039cc29c-336b-4c17-9d7d-f08488d07478\") " Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.722034 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/039cc29c-336b-4c17-9d7d-f08488d07478-kube-api-access-k5tgb" (OuterVolumeSpecName: "kube-api-access-k5tgb") pod "039cc29c-336b-4c17-9d7d-f08488d07478" (UID: "039cc29c-336b-4c17-9d7d-f08488d07478"). InnerVolumeSpecName "kube-api-access-k5tgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.755731 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "039cc29c-336b-4c17-9d7d-f08488d07478" (UID: "039cc29c-336b-4c17-9d7d-f08488d07478"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.790915 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-config-data" (OuterVolumeSpecName: "config-data") pod "039cc29c-336b-4c17-9d7d-f08488d07478" (UID: "039cc29c-336b-4c17-9d7d-f08488d07478"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.816545 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.816590 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5tgb\" (UniqueName: \"kubernetes.io/projected/039cc29c-336b-4c17-9d7d-f08488d07478-kube-api-access-k5tgb\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.816605 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/039cc29c-336b-4c17-9d7d-f08488d07478-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.927380 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-j97dl" Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.927369 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-j97dl" event={"ID":"039cc29c-336b-4c17-9d7d-f08488d07478","Type":"ContainerDied","Data":"51d01a2ae22ce9165f09c225386c0da88c7898f1ed21d7a908609446983d4023"} Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.927485 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51d01a2ae22ce9165f09c225386c0da88c7898f1ed21d7a908609446983d4023" Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.930950 4912 generic.go:334] "Generic (PLEG): container finished" podID="21bacb6a-16b3-41d4-858c-83e04de8aa9e" containerID="e4a58ad548af6a4c89df80dca47420d9345f616bc171344b3e6777ff47a6e6eb" exitCode=0 Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.932269 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" event={"ID":"21bacb6a-16b3-41d4-858c-83e04de8aa9e","Type":"ContainerDied","Data":"e4a58ad548af6a4c89df80dca47420d9345f616bc171344b3e6777ff47a6e6eb"} Dec 03 00:48:15 crc kubenswrapper[4912]: I1203 00:48:15.932316 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" event={"ID":"21bacb6a-16b3-41d4-858c-83e04de8aa9e","Type":"ContainerStarted","Data":"a15483605e04b55ab1f270221aee94d78a28410a5663e33cf0ac2eaa10219b45"} Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.443630 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-n2jc6"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.509508 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rcrbz"] Dec 03 00:48:16 crc kubenswrapper[4912]: E1203 00:48:16.510104 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="039cc29c-336b-4c17-9d7d-f08488d07478" containerName="keystone-db-sync" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.510127 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="039cc29c-336b-4c17-9d7d-f08488d07478" containerName="keystone-db-sync" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.510420 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="039cc29c-336b-4c17-9d7d-f08488d07478" containerName="keystone-db-sync" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.511300 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.525975 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.526367 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.526569 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.526691 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.527005 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rcrbz"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.527180 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-f2v25" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.539547 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-kxgcv"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.541864 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.555499 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-kxgcv"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.614373 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-kwzvj"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.615949 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-kwzvj"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.616057 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.632606 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.632905 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-7lc7t" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.640716 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-b5xlq"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.645021 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.653353 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.653784 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-frht4" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.653969 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.691732 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-svc\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.691824 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-credential-keys\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.691850 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-config\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.691872 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.691951 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-fernet-keys\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.691981 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-combined-ca-bundle\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.692009 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-scripts\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.692035 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf4zf\" (UniqueName: \"kubernetes.io/projected/762aa2d8-ebb7-497d-9667-6d58107a4489-kube-api-access-nf4zf\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.692054 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.692074 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.692110 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqphh\" (UniqueName: \"kubernetes.io/projected/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-kube-api-access-qqphh\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.692192 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-config-data\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.699530 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-b5xlq"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.748581 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-vlh4p"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.750290 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.757289 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.757492 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.757526 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qs5md" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.794720 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-svc\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.794793 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldqpn\" (UniqueName: \"kubernetes.io/projected/bf24aa53-6e61-4aae-88ed-d5085f10a073-kube-api-access-ldqpn\") pod \"heat-db-sync-kwzvj\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.794836 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-credential-keys\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.794864 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-config\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.794887 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.794930 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-combined-ca-bundle\") pod \"heat-db-sync-kwzvj\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.794953 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-combined-ca-bundle\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.794981 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-fernet-keys\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795014 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-combined-ca-bundle\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795039 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-scripts\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795068 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf4zf\" (UniqueName: \"kubernetes.io/projected/762aa2d8-ebb7-497d-9667-6d58107a4489-kube-api-access-nf4zf\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795091 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-db-sync-config-data\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795111 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795135 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-etc-machine-id\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795159 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795181 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-config-data\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795203 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-config-data\") pod \"heat-db-sync-kwzvj\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795231 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqphh\" (UniqueName: \"kubernetes.io/projected/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-kube-api-access-qqphh\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795257 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-scripts\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795321 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd4gm\" (UniqueName: \"kubernetes.io/projected/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-kube-api-access-qd4gm\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.795364 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-config-data\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.797050 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.797702 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.798364 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.799717 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-config\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.799832 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-svc\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.809187 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-combined-ca-bundle\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.813962 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-fernet-keys\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.819131 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-config-data\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.819411 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-credential-keys\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.832332 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vlh4p"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.833970 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-scripts\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.846778 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf4zf\" (UniqueName: \"kubernetes.io/projected/762aa2d8-ebb7-497d-9667-6d58107a4489-kube-api-access-nf4zf\") pod \"keystone-bootstrap-rcrbz\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.855529 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqphh\" (UniqueName: \"kubernetes.io/projected/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-kube-api-access-qqphh\") pod \"dnsmasq-dns-5b868669f-kxgcv\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.858695 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.875505 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-x5lrh"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.887996 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.894108 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901274 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd4gm\" (UniqueName: \"kubernetes.io/projected/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-kube-api-access-qd4gm\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901404 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldqpn\" (UniqueName: \"kubernetes.io/projected/bf24aa53-6e61-4aae-88ed-d5085f10a073-kube-api-access-ldqpn\") pod \"heat-db-sync-kwzvj\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901579 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-combined-ca-bundle\") pod \"heat-db-sync-kwzvj\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901623 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-combined-ca-bundle\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901685 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6g7x\" (UniqueName: \"kubernetes.io/projected/22a2b0ca-d062-446f-bcc5-6209916b6e89-kube-api-access-t6g7x\") pod \"neutron-db-sync-vlh4p\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901808 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-db-sync-config-data\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901852 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-etc-machine-id\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901876 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-combined-ca-bundle\") pod \"neutron-db-sync-vlh4p\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901907 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-config\") pod \"neutron-db-sync-vlh4p\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901928 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-config-data\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.901956 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-config-data\") pod \"heat-db-sync-kwzvj\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.902027 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-scripts\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.903030 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-etc-machine-id\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.906669 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2n742" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.907100 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-wnqwp"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.915251 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.921902 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.922267 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.922492 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wcgz7" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.934311 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-db-sync-config-data\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.935919 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.937217 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-x5lrh"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.953110 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-config-data\") pod \"heat-db-sync-kwzvj\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.956363 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-config-data\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.962704 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldqpn\" (UniqueName: \"kubernetes.io/projected/bf24aa53-6e61-4aae-88ed-d5085f10a073-kube-api-access-ldqpn\") pod \"heat-db-sync-kwzvj\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.953575 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-combined-ca-bundle\") pod \"heat-db-sync-kwzvj\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.963454 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-wnqwp"] Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.964490 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd4gm\" (UniqueName: \"kubernetes.io/projected/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-kube-api-access-qd4gm\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.965652 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-scripts\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.971236 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-kwzvj" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.976571 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-combined-ca-bundle\") pod \"cinder-db-sync-b5xlq\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.980865 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" event={"ID":"21bacb6a-16b3-41d4-858c-83e04de8aa9e","Type":"ContainerStarted","Data":"f7965bcbd8e1727a628460325f94bc010e3017cf31a0bad748076a1b407a4947"} Dec 03 00:48:16 crc kubenswrapper[4912]: I1203 00:48:16.981650 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.004862 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6g7x\" (UniqueName: \"kubernetes.io/projected/22a2b0ca-d062-446f-bcc5-6209916b6e89-kube-api-access-t6g7x\") pod \"neutron-db-sync-vlh4p\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.005019 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-db-sync-config-data\") pod \"barbican-db-sync-x5lrh\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.005222 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwfq4\" (UniqueName: \"kubernetes.io/projected/df5c7224-2a57-4931-98b4-78574fb2690b-kube-api-access-qwfq4\") pod \"barbican-db-sync-x5lrh\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.005282 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-combined-ca-bundle\") pod \"neutron-db-sync-vlh4p\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.005318 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-config\") pod \"neutron-db-sync-vlh4p\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.005496 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-combined-ca-bundle\") pod \"barbican-db-sync-x5lrh\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.016857 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-combined-ca-bundle\") pod \"neutron-db-sync-vlh4p\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.021794 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-kxgcv"] Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.022759 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-config\") pod \"neutron-db-sync-vlh4p\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.031379 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6g7x\" (UniqueName: \"kubernetes.io/projected/22a2b0ca-d062-446f-bcc5-6209916b6e89-kube-api-access-t6g7x\") pod \"neutron-db-sync-vlh4p\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.039192 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4qlxk"] Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.041226 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.052536 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.055882 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.059386 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.059583 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.068484 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4qlxk"] Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.069978 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.104573 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.108236 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwfq4\" (UniqueName: \"kubernetes.io/projected/df5c7224-2a57-4931-98b4-78574fb2690b-kube-api-access-qwfq4\") pod \"barbican-db-sync-x5lrh\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.108362 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-combined-ca-bundle\") pod \"barbican-db-sync-x5lrh\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.108517 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-scripts\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.108584 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-combined-ca-bundle\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.108624 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks5cl\" (UniqueName: \"kubernetes.io/projected/879819c5-1e33-4e31-bdfb-89d74e5efe8b-kube-api-access-ks5cl\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.108668 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/879819c5-1e33-4e31-bdfb-89d74e5efe8b-logs\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.108693 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-config-data\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.108752 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-db-sync-config-data\") pod \"barbican-db-sync-x5lrh\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.114371 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-db-sync-config-data\") pod \"barbican-db-sync-x5lrh\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.132299 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwfq4\" (UniqueName: \"kubernetes.io/projected/df5c7224-2a57-4931-98b4-78574fb2690b-kube-api-access-qwfq4\") pod \"barbican-db-sync-x5lrh\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.145530 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-combined-ca-bundle\") pod \"barbican-db-sync-x5lrh\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.149960 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" podStartSLOduration=3.149914453 podStartE2EDuration="3.149914453s" podCreationTimestamp="2025-12-03 00:48:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:48:17.01465071 +0000 UTC m=+1482.656671280" watchObservedRunningTime="2025-12-03 00:48:17.149914453 +0000 UTC m=+1482.791935013" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.232352 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.234400 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-config-data\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.245117 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97mqn\" (UniqueName: \"kubernetes.io/projected/6f2712f0-4d34-4059-b517-51da09b80ce9-kube-api-access-97mqn\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.245308 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-scripts\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.245356 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4rrc\" (UniqueName: \"kubernetes.io/projected/616696fb-abad-4f22-949a-b1fa0e87cd90-kube-api-access-p4rrc\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.245421 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-combined-ca-bundle\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.245524 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks5cl\" (UniqueName: \"kubernetes.io/projected/879819c5-1e33-4e31-bdfb-89d74e5efe8b-kube-api-access-ks5cl\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.245626 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.245688 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/879819c5-1e33-4e31-bdfb-89d74e5efe8b-logs\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.245724 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-config-data\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.245813 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-scripts\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.247843 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-run-httpd\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.247930 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.247998 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/879819c5-1e33-4e31-bdfb-89d74e5efe8b-logs\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.248037 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.248208 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.248319 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-config\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.248394 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.248576 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-log-httpd\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.255905 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-config-data\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.256892 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-scripts\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.262935 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-combined-ca-bundle\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.266196 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.274241 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks5cl\" (UniqueName: \"kubernetes.io/projected/879819c5-1e33-4e31-bdfb-89d74e5efe8b-kube-api-access-ks5cl\") pod \"placement-db-sync-wnqwp\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.328056 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.340646 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wnqwp" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.349706 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-log-httpd\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.349753 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.349782 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-config-data\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.349816 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97mqn\" (UniqueName: \"kubernetes.io/projected/6f2712f0-4d34-4059-b517-51da09b80ce9-kube-api-access-97mqn\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.349852 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4rrc\" (UniqueName: \"kubernetes.io/projected/616696fb-abad-4f22-949a-b1fa0e87cd90-kube-api-access-p4rrc\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.349896 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.349933 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-scripts\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.349959 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-run-httpd\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.349981 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.350009 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.350031 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.350056 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-config\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.350078 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.351714 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.351826 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-svc\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.352147 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-log-httpd\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.353037 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.353746 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.354190 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-run-httpd\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.354279 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-config\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.362023 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-scripts\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.363582 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.368728 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-config-data\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.369755 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.394580 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97mqn\" (UniqueName: \"kubernetes.io/projected/6f2712f0-4d34-4059-b517-51da09b80ce9-kube-api-access-97mqn\") pod \"dnsmasq-dns-cf78879c9-4qlxk\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.398760 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.416208 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4rrc\" (UniqueName: \"kubernetes.io/projected/616696fb-abad-4f22-949a-b1fa0e87cd90-kube-api-access-p4rrc\") pod \"ceilometer-0\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.424299 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:48:17 crc kubenswrapper[4912]: I1203 00:48:17.659064 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rcrbz"] Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.010525 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" podUID="21bacb6a-16b3-41d4-858c-83e04de8aa9e" containerName="dnsmasq-dns" containerID="cri-o://f7965bcbd8e1727a628460325f94bc010e3017cf31a0bad748076a1b407a4947" gracePeriod=10 Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.010734 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rcrbz" event={"ID":"762aa2d8-ebb7-497d-9667-6d58107a4489","Type":"ContainerStarted","Data":"8333b80946ea3bc5a7135c751d490c63f52bd4b48f08f12cc9136b4615cd2b3b"} Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.017766 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-kwzvj"] Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.142544 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-kxgcv"] Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.249418 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vlh4p"] Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.815172 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-x5lrh"] Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.837512 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-b5xlq"] Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.878244 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-wnqwp"] Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.884592 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4qlxk"] Dec 03 00:48:18 crc kubenswrapper[4912]: I1203 00:48:18.909999 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.028842 4912 generic.go:334] "Generic (PLEG): container finished" podID="c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" containerID="8507ba4c768effc6722ab93f8a9ec985217d6c9d0bbb06624b66d35b6c9fcc25" exitCode=0 Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.028920 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-kxgcv" event={"ID":"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37","Type":"ContainerDied","Data":"8507ba4c768effc6722ab93f8a9ec985217d6c9d0bbb06624b66d35b6c9fcc25"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.028953 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-kxgcv" event={"ID":"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37","Type":"ContainerStarted","Data":"0498b41688870f5cbd19623ea657f7f704e1aeb6427564ba742b90f55aa98a91"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.038002 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-b5xlq" event={"ID":"3071752c-9584-4ac1-8746-1b3a4ad3f4a1","Type":"ContainerStarted","Data":"eaf4e20ba560bcc1cbb1f471495ed9ffa600a1d6b4be322cfdcfb3bf73948294"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.047923 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-kwzvj" event={"ID":"bf24aa53-6e61-4aae-88ed-d5085f10a073","Type":"ContainerStarted","Data":"58de2bb02e247baabc8ff591931e1a1fafa129ba778b987f88c906d31603b1c0"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.048628 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x5lrh" event={"ID":"df5c7224-2a57-4931-98b4-78574fb2690b","Type":"ContainerStarted","Data":"cbd07aeb5461c4103bc635d06712bcd30507993223aae3aeb0358ebf52e9acc6"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.053954 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rcrbz" event={"ID":"762aa2d8-ebb7-497d-9667-6d58107a4489","Type":"ContainerStarted","Data":"234e7fbb01e07f0821d49521459c233ab76b2224b69a548795bfcd2da6441e0b"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.059451 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" event={"ID":"6f2712f0-4d34-4059-b517-51da09b80ce9","Type":"ContainerStarted","Data":"9fed1b9772fc502afdc08568599c4f526b9c9ff89e540050cc0cb73df8ad29e4"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.076674 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vlh4p" event={"ID":"22a2b0ca-d062-446f-bcc5-6209916b6e89","Type":"ContainerStarted","Data":"f85a5a5cf82b8a1257b409bbbd7e9414671a111408a649c0e4dec60ae935e4bb"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.076730 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vlh4p" event={"ID":"22a2b0ca-d062-446f-bcc5-6209916b6e89","Type":"ContainerStarted","Data":"a044de1e7e03a8afc927774679723a4d02e257564f3f8b6ed4d3ef5e1d1e8348"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.082375 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wnqwp" event={"ID":"879819c5-1e33-4e31-bdfb-89d74e5efe8b","Type":"ContainerStarted","Data":"531b722ca596a5f954ecd07e9d57531836b475038d440ab3cf8d5feda6712883"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.087568 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rcrbz" podStartSLOduration=3.087552591 podStartE2EDuration="3.087552591s" podCreationTimestamp="2025-12-03 00:48:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:48:19.080963714 +0000 UTC m=+1484.722984284" watchObservedRunningTime="2025-12-03 00:48:19.087552591 +0000 UTC m=+1484.729573141" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.109392 4912 generic.go:334] "Generic (PLEG): container finished" podID="21bacb6a-16b3-41d4-858c-83e04de8aa9e" containerID="f7965bcbd8e1727a628460325f94bc010e3017cf31a0bad748076a1b407a4947" exitCode=0 Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.109571 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" event={"ID":"21bacb6a-16b3-41d4-858c-83e04de8aa9e","Type":"ContainerDied","Data":"f7965bcbd8e1727a628460325f94bc010e3017cf31a0bad748076a1b407a4947"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.109621 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" event={"ID":"21bacb6a-16b3-41d4-858c-83e04de8aa9e","Type":"ContainerDied","Data":"a15483605e04b55ab1f270221aee94d78a28410a5663e33cf0ac2eaa10219b45"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.109634 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a15483605e04b55ab1f270221aee94d78a28410a5663e33cf0ac2eaa10219b45" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.120448 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.120637 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"616696fb-abad-4f22-949a-b1fa0e87cd90","Type":"ContainerStarted","Data":"9f9ad0422fd275be64df81cb085ddfc7fc57f707f63af227b4b0288994fc5a9f"} Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.140148 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-vlh4p" podStartSLOduration=3.140121544 podStartE2EDuration="3.140121544s" podCreationTimestamp="2025-12-03 00:48:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:48:19.107824631 +0000 UTC m=+1484.749845191" watchObservedRunningTime="2025-12-03 00:48:19.140121544 +0000 UTC m=+1484.782142104" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.233921 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-svc\") pod \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.234028 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-885df\" (UniqueName: \"kubernetes.io/projected/21bacb6a-16b3-41d4-858c-83e04de8aa9e-kube-api-access-885df\") pod \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.234246 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-nb\") pod \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.234289 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-sb\") pod \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.234317 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-config\") pod \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.234358 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-swift-storage-0\") pod \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\" (UID: \"21bacb6a-16b3-41d4-858c-83e04de8aa9e\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.272072 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21bacb6a-16b3-41d4-858c-83e04de8aa9e-kube-api-access-885df" (OuterVolumeSpecName: "kube-api-access-885df") pod "21bacb6a-16b3-41d4-858c-83e04de8aa9e" (UID: "21bacb6a-16b3-41d4-858c-83e04de8aa9e"). InnerVolumeSpecName "kube-api-access-885df". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.344665 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-885df\" (UniqueName: \"kubernetes.io/projected/21bacb6a-16b3-41d4-858c-83e04de8aa9e-kube-api-access-885df\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.435372 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "21bacb6a-16b3-41d4-858c-83e04de8aa9e" (UID: "21bacb6a-16b3-41d4-858c-83e04de8aa9e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.449317 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.579618 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-config" (OuterVolumeSpecName: "config") pod "21bacb6a-16b3-41d4-858c-83e04de8aa9e" (UID: "21bacb6a-16b3-41d4-858c-83e04de8aa9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.581770 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "21bacb6a-16b3-41d4-858c-83e04de8aa9e" (UID: "21bacb6a-16b3-41d4-858c-83e04de8aa9e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.633213 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "21bacb6a-16b3-41d4-858c-83e04de8aa9e" (UID: "21bacb6a-16b3-41d4-858c-83e04de8aa9e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.663905 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.664111 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.664223 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.704336 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "21bacb6a-16b3-41d4-858c-83e04de8aa9e" (UID: "21bacb6a-16b3-41d4-858c-83e04de8aa9e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.769748 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/21bacb6a-16b3-41d4-858c-83e04de8aa9e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.857241 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.976315 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-swift-storage-0\") pod \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.976413 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-sb\") pod \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.976554 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqphh\" (UniqueName: \"kubernetes.io/projected/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-kube-api-access-qqphh\") pod \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.976614 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-svc\") pod \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.976701 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-nb\") pod \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " Dec 03 00:48:19 crc kubenswrapper[4912]: I1203 00:48:19.976820 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-config\") pod \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\" (UID: \"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37\") " Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.029614 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-kube-api-access-qqphh" (OuterVolumeSpecName: "kube-api-access-qqphh") pod "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" (UID: "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37"). InnerVolumeSpecName "kube-api-access-qqphh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.036913 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-config" (OuterVolumeSpecName: "config") pod "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" (UID: "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.079785 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqphh\" (UniqueName: \"kubernetes.io/projected/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-kube-api-access-qqphh\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.079822 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.080166 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" (UID: "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.109001 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" (UID: "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.122967 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" (UID: "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.124985 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" (UID: "c0d3093a-1c9f-4199-b6fa-5adf4b87eb37"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.154643 4912 generic.go:334] "Generic (PLEG): container finished" podID="6f2712f0-4d34-4059-b517-51da09b80ce9" containerID="4b6094bafd4156f577df81eff79bccccb845d5ef2b8db70f115914b199486a77" exitCode=0 Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.154772 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" event={"ID":"6f2712f0-4d34-4059-b517-51da09b80ce9","Type":"ContainerDied","Data":"4b6094bafd4156f577df81eff79bccccb845d5ef2b8db70f115914b199486a77"} Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.167861 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-kxgcv" event={"ID":"c0d3093a-1c9f-4199-b6fa-5adf4b87eb37","Type":"ContainerDied","Data":"0498b41688870f5cbd19623ea657f7f704e1aeb6427564ba742b90f55aa98a91"} Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.167953 4912 scope.go:117] "RemoveContainer" containerID="8507ba4c768effc6722ab93f8a9ec985217d6c9d0bbb06624b66d35b6c9fcc25" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.168244 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-kxgcv" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.168593 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-n2jc6" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.182968 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.183007 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.183019 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.183030 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.295618 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-n2jc6"] Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.327536 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-n2jc6"] Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.372402 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-kxgcv"] Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.385518 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-kxgcv"] Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.610789 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21bacb6a-16b3-41d4-858c-83e04de8aa9e" path="/var/lib/kubelet/pods/21bacb6a-16b3-41d4-858c-83e04de8aa9e/volumes" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.611598 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" path="/var/lib/kubelet/pods/c0d3093a-1c9f-4199-b6fa-5adf4b87eb37/volumes" Dec 03 00:48:20 crc kubenswrapper[4912]: I1203 00:48:20.994863 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:48:22 crc kubenswrapper[4912]: I1203 00:48:22.305192 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" event={"ID":"6f2712f0-4d34-4059-b517-51da09b80ce9","Type":"ContainerStarted","Data":"703d3f45acd79c76aa7eafa8dd2f66e5e31432ddc2f817a46e36cf439ad37dcd"} Dec 03 00:48:22 crc kubenswrapper[4912]: I1203 00:48:22.305665 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:22 crc kubenswrapper[4912]: I1203 00:48:22.311593 4912 generic.go:334] "Generic (PLEG): container finished" podID="f78bc494-84ea-43ff-b3fc-0af2f10c43b4" containerID="2a0474e3f923d3125a4b84523cca836e5e5ba4fdfcffc3035b6dbddbbd677e31" exitCode=0 Dec 03 00:48:22 crc kubenswrapper[4912]: I1203 00:48:22.311656 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f78bc494-84ea-43ff-b3fc-0af2f10c43b4","Type":"ContainerDied","Data":"2a0474e3f923d3125a4b84523cca836e5e5ba4fdfcffc3035b6dbddbbd677e31"} Dec 03 00:48:22 crc kubenswrapper[4912]: I1203 00:48:22.348397 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" podStartSLOduration=6.348367455 podStartE2EDuration="6.348367455s" podCreationTimestamp="2025-12-03 00:48:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:48:22.327654492 +0000 UTC m=+1487.969675072" watchObservedRunningTime="2025-12-03 00:48:22.348367455 +0000 UTC m=+1487.990388015" Dec 03 00:48:23 crc kubenswrapper[4912]: I1203 00:48:23.325845 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tnx9g" event={"ID":"e760382f-ba1d-4d51-bdd0-2cd62315c4ae","Type":"ContainerStarted","Data":"8a470aa50b1dbb46c7a395ef382ffd8d3cdf3b67b5b3eb030f1e84aa26b601e7"} Dec 03 00:48:23 crc kubenswrapper[4912]: I1203 00:48:23.333823 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f78bc494-84ea-43ff-b3fc-0af2f10c43b4","Type":"ContainerStarted","Data":"67e47af29db74471b0af13eb4aa5e5e762b212a6a85af102d174ec5dd46f8ca5"} Dec 03 00:48:23 crc kubenswrapper[4912]: I1203 00:48:23.349594 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-tnx9g" podStartSLOduration=3.246671695 podStartE2EDuration="42.349570634s" podCreationTimestamp="2025-12-03 00:47:41 +0000 UTC" firstStartedPulling="2025-12-03 00:47:42.787798383 +0000 UTC m=+1448.429818943" lastFinishedPulling="2025-12-03 00:48:21.890697322 +0000 UTC m=+1487.532717882" observedRunningTime="2025-12-03 00:48:23.344053196 +0000 UTC m=+1488.986073756" watchObservedRunningTime="2025-12-03 00:48:23.349570634 +0000 UTC m=+1488.991591194" Dec 03 00:48:24 crc kubenswrapper[4912]: I1203 00:48:24.363911 4912 generic.go:334] "Generic (PLEG): container finished" podID="762aa2d8-ebb7-497d-9667-6d58107a4489" containerID="234e7fbb01e07f0821d49521459c233ab76b2224b69a548795bfcd2da6441e0b" exitCode=0 Dec 03 00:48:24 crc kubenswrapper[4912]: I1203 00:48:24.363983 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rcrbz" event={"ID":"762aa2d8-ebb7-497d-9667-6d58107a4489","Type":"ContainerDied","Data":"234e7fbb01e07f0821d49521459c233ab76b2224b69a548795bfcd2da6441e0b"} Dec 03 00:48:27 crc kubenswrapper[4912]: I1203 00:48:27.401585 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:48:27 crc kubenswrapper[4912]: I1203 00:48:27.500945 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8rk7b"] Dec 03 00:48:27 crc kubenswrapper[4912]: I1203 00:48:27.501239 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" containerID="cri-o://ff42eb77a8a089f3a34825f7746e742d10e005083c6e47ecb9b181a64c56c546" gracePeriod=10 Dec 03 00:48:28 crc kubenswrapper[4912]: I1203 00:48:28.111722 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: connect: connection refused" Dec 03 00:48:28 crc kubenswrapper[4912]: I1203 00:48:28.418694 4912 generic.go:334] "Generic (PLEG): container finished" podID="924dacce-93b0-4492-bded-aa0ac2668954" containerID="ff42eb77a8a089f3a34825f7746e742d10e005083c6e47ecb9b181a64c56c546" exitCode=0 Dec 03 00:48:28 crc kubenswrapper[4912]: I1203 00:48:28.418772 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" event={"ID":"924dacce-93b0-4492-bded-aa0ac2668954","Type":"ContainerDied","Data":"ff42eb77a8a089f3a34825f7746e742d10e005083c6e47ecb9b181a64c56c546"} Dec 03 00:48:29 crc kubenswrapper[4912]: I1203 00:48:29.435252 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f78bc494-84ea-43ff-b3fc-0af2f10c43b4","Type":"ContainerStarted","Data":"582b07eb242f3bc49cd5edd290a92ec653e18a92801efde65871d9827373c51b"} Dec 03 00:48:33 crc kubenswrapper[4912]: I1203 00:48:33.113342 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: connect: connection refused" Dec 03 00:48:35 crc kubenswrapper[4912]: E1203 00:48:35.772451 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 03 00:48:35 crc kubenswrapper[4912]: E1203 00:48:35.773596 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwfq4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-x5lrh_openstack(df5c7224-2a57-4931-98b4-78574fb2690b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:48:35 crc kubenswrapper[4912]: E1203 00:48:35.775110 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-x5lrh" podUID="df5c7224-2a57-4931-98b4-78574fb2690b" Dec 03 00:48:35 crc kubenswrapper[4912]: I1203 00:48:35.908049 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.080386 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-combined-ca-bundle\") pod \"762aa2d8-ebb7-497d-9667-6d58107a4489\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.081963 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-config-data\") pod \"762aa2d8-ebb7-497d-9667-6d58107a4489\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.082120 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-credential-keys\") pod \"762aa2d8-ebb7-497d-9667-6d58107a4489\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.082149 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-fernet-keys\") pod \"762aa2d8-ebb7-497d-9667-6d58107a4489\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.082198 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-scripts\") pod \"762aa2d8-ebb7-497d-9667-6d58107a4489\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.082226 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf4zf\" (UniqueName: \"kubernetes.io/projected/762aa2d8-ebb7-497d-9667-6d58107a4489-kube-api-access-nf4zf\") pod \"762aa2d8-ebb7-497d-9667-6d58107a4489\" (UID: \"762aa2d8-ebb7-497d-9667-6d58107a4489\") " Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.090317 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-scripts" (OuterVolumeSpecName: "scripts") pod "762aa2d8-ebb7-497d-9667-6d58107a4489" (UID: "762aa2d8-ebb7-497d-9667-6d58107a4489"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.090337 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "762aa2d8-ebb7-497d-9667-6d58107a4489" (UID: "762aa2d8-ebb7-497d-9667-6d58107a4489"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.090488 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/762aa2d8-ebb7-497d-9667-6d58107a4489-kube-api-access-nf4zf" (OuterVolumeSpecName: "kube-api-access-nf4zf") pod "762aa2d8-ebb7-497d-9667-6d58107a4489" (UID: "762aa2d8-ebb7-497d-9667-6d58107a4489"). InnerVolumeSpecName "kube-api-access-nf4zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.117071 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "762aa2d8-ebb7-497d-9667-6d58107a4489" (UID: "762aa2d8-ebb7-497d-9667-6d58107a4489"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.126618 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "762aa2d8-ebb7-497d-9667-6d58107a4489" (UID: "762aa2d8-ebb7-497d-9667-6d58107a4489"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.134898 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-config-data" (OuterVolumeSpecName: "config-data") pod "762aa2d8-ebb7-497d-9667-6d58107a4489" (UID: "762aa2d8-ebb7-497d-9667-6d58107a4489"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.184931 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.184977 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.184987 4912 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.184997 4912 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.185007 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/762aa2d8-ebb7-497d-9667-6d58107a4489-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.185042 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf4zf\" (UniqueName: \"kubernetes.io/projected/762aa2d8-ebb7-497d-9667-6d58107a4489-kube-api-access-nf4zf\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.596046 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rcrbz" Dec 03 00:48:36 crc kubenswrapper[4912]: E1203 00:48:36.598885 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-x5lrh" podUID="df5c7224-2a57-4931-98b4-78574fb2690b" Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.602194 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rcrbz" event={"ID":"762aa2d8-ebb7-497d-9667-6d58107a4489","Type":"ContainerDied","Data":"8333b80946ea3bc5a7135c751d490c63f52bd4b48f08f12cc9136b4615cd2b3b"} Dec 03 00:48:36 crc kubenswrapper[4912]: I1203 00:48:36.602261 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8333b80946ea3bc5a7135c751d490c63f52bd4b48f08f12cc9136b4615cd2b3b" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.113079 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rcrbz"] Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.123880 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rcrbz"] Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.208041 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dkrd4"] Dec 03 00:48:37 crc kubenswrapper[4912]: E1203 00:48:37.208502 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21bacb6a-16b3-41d4-858c-83e04de8aa9e" containerName="dnsmasq-dns" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.208521 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="21bacb6a-16b3-41d4-858c-83e04de8aa9e" containerName="dnsmasq-dns" Dec 03 00:48:37 crc kubenswrapper[4912]: E1203 00:48:37.208559 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" containerName="init" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.208566 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" containerName="init" Dec 03 00:48:37 crc kubenswrapper[4912]: E1203 00:48:37.208583 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="762aa2d8-ebb7-497d-9667-6d58107a4489" containerName="keystone-bootstrap" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.208591 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="762aa2d8-ebb7-497d-9667-6d58107a4489" containerName="keystone-bootstrap" Dec 03 00:48:37 crc kubenswrapper[4912]: E1203 00:48:37.208600 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21bacb6a-16b3-41d4-858c-83e04de8aa9e" containerName="init" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.208606 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="21bacb6a-16b3-41d4-858c-83e04de8aa9e" containerName="init" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.208834 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="762aa2d8-ebb7-497d-9667-6d58107a4489" containerName="keystone-bootstrap" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.208855 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="21bacb6a-16b3-41d4-858c-83e04de8aa9e" containerName="dnsmasq-dns" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.208871 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0d3093a-1c9f-4199-b6fa-5adf4b87eb37" containerName="init" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.209670 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.212577 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.213216 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.213356 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.213594 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-f2v25" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.213610 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.215562 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r248j\" (UniqueName: \"kubernetes.io/projected/d4bd0e42-218d-47de-a16f-214d5f4d2910-kube-api-access-r248j\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.215636 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-credential-keys\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.215847 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-scripts\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.215977 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-config-data\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.216068 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-combined-ca-bundle\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.216189 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-fernet-keys\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.238781 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dkrd4"] Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.322589 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r248j\" (UniqueName: \"kubernetes.io/projected/d4bd0e42-218d-47de-a16f-214d5f4d2910-kube-api-access-r248j\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.323195 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-credential-keys\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.324422 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-scripts\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.324597 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-config-data\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.324705 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-combined-ca-bundle\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.324800 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-fernet-keys\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.328823 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-credential-keys\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.335457 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-scripts\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.335663 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-fernet-keys\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.335754 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-config-data\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.335929 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-combined-ca-bundle\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.343246 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r248j\" (UniqueName: \"kubernetes.io/projected/d4bd0e42-218d-47de-a16f-214d5f4d2910-kube-api-access-r248j\") pod \"keystone-bootstrap-dkrd4\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:37 crc kubenswrapper[4912]: I1203 00:48:37.539307 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:48:38 crc kubenswrapper[4912]: I1203 00:48:38.590328 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="762aa2d8-ebb7-497d-9667-6d58107a4489" path="/var/lib/kubelet/pods/762aa2d8-ebb7-497d-9667-6d58107a4489/volumes" Dec 03 00:48:43 crc kubenswrapper[4912]: I1203 00:48:43.136062 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: i/o timeout" Dec 03 00:48:43 crc kubenswrapper[4912]: I1203 00:48:43.137288 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:48:44 crc kubenswrapper[4912]: I1203 00:48:44.703689 4912 generic.go:334] "Generic (PLEG): container finished" podID="e760382f-ba1d-4d51-bdd0-2cd62315c4ae" containerID="8a470aa50b1dbb46c7a395ef382ffd8d3cdf3b67b5b3eb030f1e84aa26b601e7" exitCode=0 Dec 03 00:48:44 crc kubenswrapper[4912]: I1203 00:48:44.703752 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tnx9g" event={"ID":"e760382f-ba1d-4d51-bdd0-2cd62315c4ae","Type":"ContainerDied","Data":"8a470aa50b1dbb46c7a395ef382ffd8d3cdf3b67b5b3eb030f1e84aa26b601e7"} Dec 03 00:48:46 crc kubenswrapper[4912]: I1203 00:48:46.727986 4912 generic.go:334] "Generic (PLEG): container finished" podID="22a2b0ca-d062-446f-bcc5-6209916b6e89" containerID="f85a5a5cf82b8a1257b409bbbd7e9414671a111408a649c0e4dec60ae935e4bb" exitCode=0 Dec 03 00:48:46 crc kubenswrapper[4912]: I1203 00:48:46.728084 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vlh4p" event={"ID":"22a2b0ca-d062-446f-bcc5-6209916b6e89","Type":"ContainerDied","Data":"f85a5a5cf82b8a1257b409bbbd7e9414671a111408a649c0e4dec60ae935e4bb"} Dec 03 00:48:48 crc kubenswrapper[4912]: I1203 00:48:48.083006 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:48:48 crc kubenswrapper[4912]: I1203 00:48:48.083660 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:48:48 crc kubenswrapper[4912]: I1203 00:48:48.137031 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: i/o timeout" Dec 03 00:48:53 crc kubenswrapper[4912]: I1203 00:48:53.137769 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: i/o timeout" Dec 03 00:48:54 crc kubenswrapper[4912]: E1203 00:48:54.368832 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 03 00:48:54 crc kubenswrapper[4912]: E1203 00:48:54.369123 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd4h668h66fh66h5c5h646h667h589h584h5ddh5fh659h668h9fh54dhd7h64fh577h5b4h596hf8h88h554h658h57ch86hfdh5cch585h65fhcbhd4q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p4rrc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(616696fb-abad-4f22-949a-b1fa0e87cd90): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.510544 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.547729 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tnx9g" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.551087 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.616660 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-config-data\") pod \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.616777 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-config\") pod \"22a2b0ca-d062-446f-bcc5-6209916b6e89\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.616902 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-combined-ca-bundle\") pod \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.616926 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-sb\") pod \"924dacce-93b0-4492-bded-aa0ac2668954\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.616948 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65php\" (UniqueName: \"kubernetes.io/projected/924dacce-93b0-4492-bded-aa0ac2668954-kube-api-access-65php\") pod \"924dacce-93b0-4492-bded-aa0ac2668954\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.616980 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-dns-svc\") pod \"924dacce-93b0-4492-bded-aa0ac2668954\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.617031 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6g7x\" (UniqueName: \"kubernetes.io/projected/22a2b0ca-d062-446f-bcc5-6209916b6e89-kube-api-access-t6g7x\") pod \"22a2b0ca-d062-446f-bcc5-6209916b6e89\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.617058 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk7cz\" (UniqueName: \"kubernetes.io/projected/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-kube-api-access-mk7cz\") pod \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.617141 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-combined-ca-bundle\") pod \"22a2b0ca-d062-446f-bcc5-6209916b6e89\" (UID: \"22a2b0ca-d062-446f-bcc5-6209916b6e89\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.617181 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-config\") pod \"924dacce-93b0-4492-bded-aa0ac2668954\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.617289 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-db-sync-config-data\") pod \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\" (UID: \"e760382f-ba1d-4d51-bdd0-2cd62315c4ae\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.617311 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-nb\") pod \"924dacce-93b0-4492-bded-aa0ac2668954\" (UID: \"924dacce-93b0-4492-bded-aa0ac2668954\") " Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.674731 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e760382f-ba1d-4d51-bdd0-2cd62315c4ae" (UID: "e760382f-ba1d-4d51-bdd0-2cd62315c4ae"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.675515 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22a2b0ca-d062-446f-bcc5-6209916b6e89-kube-api-access-t6g7x" (OuterVolumeSpecName: "kube-api-access-t6g7x") pod "22a2b0ca-d062-446f-bcc5-6209916b6e89" (UID: "22a2b0ca-d062-446f-bcc5-6209916b6e89"). InnerVolumeSpecName "kube-api-access-t6g7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.691758 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/924dacce-93b0-4492-bded-aa0ac2668954-kube-api-access-65php" (OuterVolumeSpecName: "kube-api-access-65php") pod "924dacce-93b0-4492-bded-aa0ac2668954" (UID: "924dacce-93b0-4492-bded-aa0ac2668954"). InnerVolumeSpecName "kube-api-access-65php". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.701697 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-kube-api-access-mk7cz" (OuterVolumeSpecName: "kube-api-access-mk7cz") pod "e760382f-ba1d-4d51-bdd0-2cd62315c4ae" (UID: "e760382f-ba1d-4d51-bdd0-2cd62315c4ae"). InnerVolumeSpecName "kube-api-access-mk7cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.720349 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65php\" (UniqueName: \"kubernetes.io/projected/924dacce-93b0-4492-bded-aa0ac2668954-kube-api-access-65php\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.720400 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6g7x\" (UniqueName: \"kubernetes.io/projected/22a2b0ca-d062-446f-bcc5-6209916b6e89-kube-api-access-t6g7x\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.720411 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk7cz\" (UniqueName: \"kubernetes.io/projected/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-kube-api-access-mk7cz\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.720421 4912 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.772625 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-config" (OuterVolumeSpecName: "config") pod "22a2b0ca-d062-446f-bcc5-6209916b6e89" (UID: "22a2b0ca-d062-446f-bcc5-6209916b6e89"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.782306 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e760382f-ba1d-4d51-bdd0-2cd62315c4ae" (UID: "e760382f-ba1d-4d51-bdd0-2cd62315c4ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.797295 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "924dacce-93b0-4492-bded-aa0ac2668954" (UID: "924dacce-93b0-4492-bded-aa0ac2668954"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.803217 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "924dacce-93b0-4492-bded-aa0ac2668954" (UID: "924dacce-93b0-4492-bded-aa0ac2668954"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.805824 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "924dacce-93b0-4492-bded-aa0ac2668954" (UID: "924dacce-93b0-4492-bded-aa0ac2668954"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.806098 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22a2b0ca-d062-446f-bcc5-6209916b6e89" (UID: "22a2b0ca-d062-446f-bcc5-6209916b6e89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.823198 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.823225 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.823238 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.823248 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.823257 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.823272 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/22a2b0ca-d062-446f-bcc5-6209916b6e89-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.827771 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-config" (OuterVolumeSpecName: "config") pod "924dacce-93b0-4492-bded-aa0ac2668954" (UID: "924dacce-93b0-4492-bded-aa0ac2668954"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.828132 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-config-data" (OuterVolumeSpecName: "config-data") pod "e760382f-ba1d-4d51-bdd0-2cd62315c4ae" (UID: "e760382f-ba1d-4d51-bdd0-2cd62315c4ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.832702 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" event={"ID":"924dacce-93b0-4492-bded-aa0ac2668954","Type":"ContainerDied","Data":"d3e93ff36465134334c09c375a47cdb574dc1b3b52caa4c6977ece9016b6c1b8"} Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.832785 4912 scope.go:117] "RemoveContainer" containerID="ff42eb77a8a089f3a34825f7746e742d10e005083c6e47ecb9b181a64c56c546" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.832888 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.837375 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vlh4p" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.837389 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vlh4p" event={"ID":"22a2b0ca-d062-446f-bcc5-6209916b6e89","Type":"ContainerDied","Data":"a044de1e7e03a8afc927774679723a4d02e257564f3f8b6ed4d3ef5e1d1e8348"} Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.837531 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a044de1e7e03a8afc927774679723a4d02e257564f3f8b6ed4d3ef5e1d1e8348" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.843525 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tnx9g" event={"ID":"e760382f-ba1d-4d51-bdd0-2cd62315c4ae","Type":"ContainerDied","Data":"c7fca989ce4af296a93885a099803984be7d9f5ef3dd171ba02d35bd74f284a5"} Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.843569 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7fca989ce4af296a93885a099803984be7d9f5ef3dd171ba02d35bd74f284a5" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.843636 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tnx9g" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.882903 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8rk7b"] Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.892759 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-8rk7b"] Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.926957 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e760382f-ba1d-4d51-bdd0-2cd62315c4ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:54 crc kubenswrapper[4912]: I1203 00:48:54.927006 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/924dacce-93b0-4492-bded-aa0ac2668954-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:48:55 crc kubenswrapper[4912]: E1203 00:48:55.073255 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 03 00:48:55 crc kubenswrapper[4912]: E1203 00:48:55.073493 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ldqpn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-kwzvj_openstack(bf24aa53-6e61-4aae-88ed-d5085f10a073): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:48:55 crc kubenswrapper[4912]: E1203 00:48:55.074743 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-kwzvj" podUID="bf24aa53-6e61-4aae-88ed-d5085f10a073" Dec 03 00:48:55 crc kubenswrapper[4912]: E1203 00:48:55.878186 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-kwzvj" podUID="bf24aa53-6e61-4aae-88ed-d5085f10a073" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.906483 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79cd4f6685-2b8n5"] Dec 03 00:48:55 crc kubenswrapper[4912]: E1203 00:48:55.907081 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.907102 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" Dec 03 00:48:55 crc kubenswrapper[4912]: E1203 00:48:55.907129 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a2b0ca-d062-446f-bcc5-6209916b6e89" containerName="neutron-db-sync" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.907136 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a2b0ca-d062-446f-bcc5-6209916b6e89" containerName="neutron-db-sync" Dec 03 00:48:55 crc kubenswrapper[4912]: E1203 00:48:55.907154 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="init" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.907161 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="init" Dec 03 00:48:55 crc kubenswrapper[4912]: E1203 00:48:55.907175 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e760382f-ba1d-4d51-bdd0-2cd62315c4ae" containerName="glance-db-sync" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.907182 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e760382f-ba1d-4d51-bdd0-2cd62315c4ae" containerName="glance-db-sync" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.907421 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="e760382f-ba1d-4d51-bdd0-2cd62315c4ae" containerName="glance-db-sync" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.907475 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="22a2b0ca-d062-446f-bcc5-6209916b6e89" containerName="neutron-db-sync" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.907496 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.908900 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.938560 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79cd4f6685-2b8n5"] Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.997727 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhz8w\" (UniqueName: \"kubernetes.io/projected/ed75d2d6-0541-4c1d-b421-ec41712f2993-kube-api-access-rhz8w\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.997782 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-config\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.997819 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-swift-storage-0\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.997878 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-svc\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.997959 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-nb\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:55 crc kubenswrapper[4912]: I1203 00:48:55.998058 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-sb\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.101649 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-sb\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.101747 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhz8w\" (UniqueName: \"kubernetes.io/projected/ed75d2d6-0541-4c1d-b421-ec41712f2993-kube-api-access-rhz8w\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.101778 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-config\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.101814 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-swift-storage-0\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.101870 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-svc\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.101977 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-nb\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.103117 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-nb\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.103689 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-config\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.103704 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-sb\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.109338 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-swift-storage-0\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.109505 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-svc\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.148518 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhz8w\" (UniqueName: \"kubernetes.io/projected/ed75d2d6-0541-4c1d-b421-ec41712f2993-kube-api-access-rhz8w\") pod \"dnsmasq-dns-79cd4f6685-2b8n5\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.150356 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-54f689d9c8-bxnjz"] Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.152519 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.164875 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.165237 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.165383 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qs5md" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.165604 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.202823 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54f689d9c8-bxnjz"] Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.264942 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.279068 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79cd4f6685-2b8n5"] Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.307084 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76whb\" (UniqueName: \"kubernetes.io/projected/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-kube-api-access-76whb\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.307144 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-combined-ca-bundle\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.307194 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-ovndb-tls-certs\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.307293 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-config\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.307328 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-httpd-config\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.322601 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6b98f"] Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.328324 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.369060 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6b98f"] Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.409754 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-ovndb-tls-certs\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410593 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f24zm\" (UniqueName: \"kubernetes.io/projected/dc3a3200-2b91-41bb-8a80-25a8d2601331-kube-api-access-f24zm\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410692 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-config\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410720 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-config\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410740 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410777 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410808 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-httpd-config\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410825 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-svc\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410914 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76whb\" (UniqueName: \"kubernetes.io/projected/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-kube-api-access-76whb\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410945 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-combined-ca-bundle\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.410965 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.416191 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-ovndb-tls-certs\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.417294 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-config\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.419179 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-httpd-config\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.419820 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-combined-ca-bundle\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.432000 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76whb\" (UniqueName: \"kubernetes.io/projected/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-kube-api-access-76whb\") pod \"neutron-54f689d9c8-bxnjz\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.516130 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.516230 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f24zm\" (UniqueName: \"kubernetes.io/projected/dc3a3200-2b91-41bb-8a80-25a8d2601331-kube-api-access-f24zm\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.516287 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-config\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.516309 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.516339 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.516362 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-svc\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.517226 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-svc\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.517744 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.523752 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.526962 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-config\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.529057 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.541569 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f24zm\" (UniqueName: \"kubernetes.io/projected/dc3a3200-2b91-41bb-8a80-25a8d2601331-kube-api-access-f24zm\") pod \"dnsmasq-dns-6b7b667979-6b98f\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.542082 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.596094 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="924dacce-93b0-4492-bded-aa0ac2668954" path="/var/lib/kubelet/pods/924dacce-93b0-4492-bded-aa0ac2668954/volumes" Dec 03 00:48:56 crc kubenswrapper[4912]: I1203 00:48:56.667923 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:48:57 crc kubenswrapper[4912]: I1203 00:48:57.482215 4912 scope.go:117] "RemoveContainer" containerID="85331a75effc382414de473f18aa352fb989a2da8d27a27c1f8d5b8c4c82e321" Dec 03 00:48:57 crc kubenswrapper[4912]: E1203 00:48:57.495488 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 00:48:57 crc kubenswrapper[4912]: E1203 00:48:57.495723 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qd4gm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-b5xlq_openstack(3071752c-9584-4ac1-8746-1b3a4ad3f4a1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:48:57 crc kubenswrapper[4912]: E1203 00:48:57.497654 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-b5xlq" podUID="3071752c-9584-4ac1-8746-1b3a4ad3f4a1" Dec 03 00:48:57 crc kubenswrapper[4912]: E1203 00:48:57.929496 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-b5xlq" podUID="3071752c-9584-4ac1-8746-1b3a4ad3f4a1" Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.057993 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dkrd4"] Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.139865 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-8rk7b" podUID="924dacce-93b0-4492-bded-aa0ac2668954" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: i/o timeout" Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.349805 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79cd4f6685-2b8n5"] Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.447000 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54f689d9c8-bxnjz"] Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.525828 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6b98f"] Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.929059 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" event={"ID":"ed75d2d6-0541-4c1d-b421-ec41712f2993","Type":"ContainerStarted","Data":"c70ef3df56992f925c6ae4fa37f807dc49d25ee29d8bee77e62b87bbbf15cff2"} Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.932337 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wnqwp" event={"ID":"879819c5-1e33-4e31-bdfb-89d74e5efe8b","Type":"ContainerStarted","Data":"c8b3de5502e8daf8ff011cc14855b97c41f69269c1f76f6835a9e3ddba99d0d9"} Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.941187 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dkrd4" event={"ID":"d4bd0e42-218d-47de-a16f-214d5f4d2910","Type":"ContainerStarted","Data":"10da5eb5c055fa52db79e62a49ca41e7146cd4d94cdcc498e5edc7b6ceb3a6c6"} Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.941283 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dkrd4" event={"ID":"d4bd0e42-218d-47de-a16f-214d5f4d2910","Type":"ContainerStarted","Data":"c641c2eb8b20224b86ceb7e079a04b6b4248be8a4134f48cd5fee945cc6093bf"} Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.958214 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-wnqwp" podStartSLOduration=6.859008133 podStartE2EDuration="42.958184673s" podCreationTimestamp="2025-12-03 00:48:16 +0000 UTC" firstStartedPulling="2025-12-03 00:48:18.951798935 +0000 UTC m=+1484.593819495" lastFinishedPulling="2025-12-03 00:48:55.050975475 +0000 UTC m=+1520.692996035" observedRunningTime="2025-12-03 00:48:58.957483504 +0000 UTC m=+1524.599504074" watchObservedRunningTime="2025-12-03 00:48:58.958184673 +0000 UTC m=+1524.600205233" Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.962865 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"f78bc494-84ea-43ff-b3fc-0af2f10c43b4","Type":"ContainerStarted","Data":"02c679e3f025895e5e05f13183798335ab7ce560968e736ba1ff80f413146257"} Dec 03 00:48:58 crc kubenswrapper[4912]: I1203 00:48:58.984848 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x5lrh" event={"ID":"df5c7224-2a57-4931-98b4-78574fb2690b","Type":"ContainerStarted","Data":"ac4118f51a831c59975ddac5999c44968efe58969edd7422dfc30604341e260d"} Dec 03 00:48:59 crc kubenswrapper[4912]: I1203 00:48:59.013250 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dkrd4" podStartSLOduration=22.013220793 podStartE2EDuration="22.013220793s" podCreationTimestamp="2025-12-03 00:48:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:48:58.98463055 +0000 UTC m=+1524.626651120" watchObservedRunningTime="2025-12-03 00:48:59.013220793 +0000 UTC m=+1524.655241353" Dec 03 00:48:59 crc kubenswrapper[4912]: I1203 00:48:59.046723 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=50.046698067 podStartE2EDuration="50.046698067s" podCreationTimestamp="2025-12-03 00:48:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:48:59.023909838 +0000 UTC m=+1524.665930418" watchObservedRunningTime="2025-12-03 00:48:59.046698067 +0000 UTC m=+1524.688718627" Dec 03 00:48:59 crc kubenswrapper[4912]: I1203 00:48:59.057581 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-x5lrh" podStartSLOduration=4.273597335 podStartE2EDuration="43.057557717s" podCreationTimestamp="2025-12-03 00:48:16 +0000 UTC" firstStartedPulling="2025-12-03 00:48:18.90594663 +0000 UTC m=+1484.547967190" lastFinishedPulling="2025-12-03 00:48:57.689907012 +0000 UTC m=+1523.331927572" observedRunningTime="2025-12-03 00:48:59.040577604 +0000 UTC m=+1524.682598164" watchObservedRunningTime="2025-12-03 00:48:59.057557717 +0000 UTC m=+1524.699578277" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.030736 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f689d9c8-bxnjz" event={"ID":"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3","Type":"ContainerStarted","Data":"a8ae68d8961560dd4cb9207fc6ae82494366831638580b903327485478be6d64"} Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.031588 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f689d9c8-bxnjz" event={"ID":"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3","Type":"ContainerStarted","Data":"434b65e23d30e8f8fbc3e6d469fc7d72f358238434a252dc6346edb0bd0c4ac4"} Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.046591 4912 generic.go:334] "Generic (PLEG): container finished" podID="ed75d2d6-0541-4c1d-b421-ec41712f2993" containerID="e8130333d03c3b7814583caf20b0c6fa0d281699eb817a6fe5fe7767b77aae46" exitCode=0 Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.046849 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" event={"ID":"ed75d2d6-0541-4c1d-b421-ec41712f2993","Type":"ContainerDied","Data":"e8130333d03c3b7814583caf20b0c6fa0d281699eb817a6fe5fe7767b77aae46"} Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.056060 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" event={"ID":"dc3a3200-2b91-41bb-8a80-25a8d2601331","Type":"ContainerStarted","Data":"4a157c5e14db42e4d5f3cb81957deeeb118b425f226e6621ee24676bed2a85d2"} Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.056115 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" event={"ID":"dc3a3200-2b91-41bb-8a80-25a8d2601331","Type":"ContainerStarted","Data":"042c550ebe94a08a8485d5caeea879862fac6b6ef1d5325b9f6c7888c97516b8"} Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.063489 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"616696fb-abad-4f22-949a-b1fa0e87cd90","Type":"ContainerStarted","Data":"2a585e40cc6f95d08d378351d5a1ca2a190d4059b133042e0dbe2586ddfef79b"} Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.392753 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.598927 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.742276 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-config\") pod \"ed75d2d6-0541-4c1d-b421-ec41712f2993\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.742347 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-svc\") pod \"ed75d2d6-0541-4c1d-b421-ec41712f2993\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.742442 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-sb\") pod \"ed75d2d6-0541-4c1d-b421-ec41712f2993\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.742469 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-nb\") pod \"ed75d2d6-0541-4c1d-b421-ec41712f2993\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.742551 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-swift-storage-0\") pod \"ed75d2d6-0541-4c1d-b421-ec41712f2993\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.742786 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhz8w\" (UniqueName: \"kubernetes.io/projected/ed75d2d6-0541-4c1d-b421-ec41712f2993-kube-api-access-rhz8w\") pod \"ed75d2d6-0541-4c1d-b421-ec41712f2993\" (UID: \"ed75d2d6-0541-4c1d-b421-ec41712f2993\") " Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.776153 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed75d2d6-0541-4c1d-b421-ec41712f2993-kube-api-access-rhz8w" (OuterVolumeSpecName: "kube-api-access-rhz8w") pod "ed75d2d6-0541-4c1d-b421-ec41712f2993" (UID: "ed75d2d6-0541-4c1d-b421-ec41712f2993"). InnerVolumeSpecName "kube-api-access-rhz8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.777448 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ed75d2d6-0541-4c1d-b421-ec41712f2993" (UID: "ed75d2d6-0541-4c1d-b421-ec41712f2993"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.793155 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ed75d2d6-0541-4c1d-b421-ec41712f2993" (UID: "ed75d2d6-0541-4c1d-b421-ec41712f2993"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.795919 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-config" (OuterVolumeSpecName: "config") pod "ed75d2d6-0541-4c1d-b421-ec41712f2993" (UID: "ed75d2d6-0541-4c1d-b421-ec41712f2993"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.800002 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ed75d2d6-0541-4c1d-b421-ec41712f2993" (UID: "ed75d2d6-0541-4c1d-b421-ec41712f2993"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.845610 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.845653 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.845666 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.845678 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhz8w\" (UniqueName: \"kubernetes.io/projected/ed75d2d6-0541-4c1d-b421-ec41712f2993-kube-api-access-rhz8w\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.845688 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.856134 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ed75d2d6-0541-4c1d-b421-ec41712f2993" (UID: "ed75d2d6-0541-4c1d-b421-ec41712f2993"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:00 crc kubenswrapper[4912]: I1203 00:49:00.950301 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ed75d2d6-0541-4c1d-b421-ec41712f2993-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.083316 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" event={"ID":"ed75d2d6-0541-4c1d-b421-ec41712f2993","Type":"ContainerDied","Data":"c70ef3df56992f925c6ae4fa37f807dc49d25ee29d8bee77e62b87bbbf15cff2"} Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.083391 4912 scope.go:117] "RemoveContainer" containerID="e8130333d03c3b7814583caf20b0c6fa0d281699eb817a6fe5fe7767b77aae46" Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.083613 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79cd4f6685-2b8n5" Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.093040 4912 generic.go:334] "Generic (PLEG): container finished" podID="dc3a3200-2b91-41bb-8a80-25a8d2601331" containerID="4a157c5e14db42e4d5f3cb81957deeeb118b425f226e6621ee24676bed2a85d2" exitCode=0 Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.093119 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" event={"ID":"dc3a3200-2b91-41bb-8a80-25a8d2601331","Type":"ContainerDied","Data":"4a157c5e14db42e4d5f3cb81957deeeb118b425f226e6621ee24676bed2a85d2"} Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.093215 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" event={"ID":"dc3a3200-2b91-41bb-8a80-25a8d2601331","Type":"ContainerStarted","Data":"c02d91c0179dfcf5c01d4d143c4d8a4bf3c0327e0171a843603c1402cdd86248"} Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.093310 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.103505 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f689d9c8-bxnjz" event={"ID":"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3","Type":"ContainerStarted","Data":"9f1699b7d36aca8216a0caf12a1d18f162ef66ab271b63e30b8ddea67b51299f"} Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.103790 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.119219 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" podStartSLOduration=5.119192055 podStartE2EDuration="5.119192055s" podCreationTimestamp="2025-12-03 00:48:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:01.115384044 +0000 UTC m=+1526.757404614" watchObservedRunningTime="2025-12-03 00:49:01.119192055 +0000 UTC m=+1526.761212625" Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.212116 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79cd4f6685-2b8n5"] Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.232503 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79cd4f6685-2b8n5"] Dec 03 00:49:01 crc kubenswrapper[4912]: I1203 00:49:01.240620 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-54f689d9c8-bxnjz" podStartSLOduration=5.240594448 podStartE2EDuration="5.240594448s" podCreationTimestamp="2025-12-03 00:48:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:01.210016461 +0000 UTC m=+1526.852037021" watchObservedRunningTime="2025-12-03 00:49:01.240594448 +0000 UTC m=+1526.882615008" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.131403 4912 generic.go:334] "Generic (PLEG): container finished" podID="879819c5-1e33-4e31-bdfb-89d74e5efe8b" containerID="c8b3de5502e8daf8ff011cc14855b97c41f69269c1f76f6835a9e3ddba99d0d9" exitCode=0 Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.131479 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wnqwp" event={"ID":"879819c5-1e33-4e31-bdfb-89d74e5efe8b","Type":"ContainerDied","Data":"c8b3de5502e8daf8ff011cc14855b97c41f69269c1f76f6835a9e3ddba99d0d9"} Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.496185 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-59fd659dd5-vmcdc"] Dec 03 00:49:02 crc kubenswrapper[4912]: E1203 00:49:02.496779 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed75d2d6-0541-4c1d-b421-ec41712f2993" containerName="init" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.496801 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed75d2d6-0541-4c1d-b421-ec41712f2993" containerName="init" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.497068 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed75d2d6-0541-4c1d-b421-ec41712f2993" containerName="init" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.498311 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.511341 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.511362 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.521784 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59fd659dd5-vmcdc"] Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.616188 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-config\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.616263 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-httpd-config\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.616356 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-ovndb-tls-certs\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.616378 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-public-tls-certs\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.616459 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-combined-ca-bundle\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.616524 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbthn\" (UniqueName: \"kubernetes.io/projected/6b4100ae-eb0e-402b-b442-d2d313e2bff4-kube-api-access-gbthn\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.616582 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-internal-tls-certs\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.655121 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed75d2d6-0541-4c1d-b421-ec41712f2993" path="/var/lib/kubelet/pods/ed75d2d6-0541-4c1d-b421-ec41712f2993/volumes" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.719212 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-config\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.719265 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-httpd-config\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.719347 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-ovndb-tls-certs\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.719370 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-public-tls-certs\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.719446 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-combined-ca-bundle\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.719494 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbthn\" (UniqueName: \"kubernetes.io/projected/6b4100ae-eb0e-402b-b442-d2d313e2bff4-kube-api-access-gbthn\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.719569 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-internal-tls-certs\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.730417 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-internal-tls-certs\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.735259 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-public-tls-certs\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.735857 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-config\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.744566 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-ovndb-tls-certs\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.746022 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-combined-ca-bundle\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.750308 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbthn\" (UniqueName: \"kubernetes.io/projected/6b4100ae-eb0e-402b-b442-d2d313e2bff4-kube-api-access-gbthn\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.751856 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6b4100ae-eb0e-402b-b442-d2d313e2bff4-httpd-config\") pod \"neutron-59fd659dd5-vmcdc\" (UID: \"6b4100ae-eb0e-402b-b442-d2d313e2bff4\") " pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:02 crc kubenswrapper[4912]: I1203 00:49:02.834212 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.591502 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wnqwp" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.608800 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59fd659dd5-vmcdc"] Dec 03 00:49:03 crc kubenswrapper[4912]: W1203 00:49:03.634330 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b4100ae_eb0e_402b_b442_d2d313e2bff4.slice/crio-96fb9fd67b99b130781db48e3e1c9e27fd9aea890727c2dc636511547eb9c736 WatchSource:0}: Error finding container 96fb9fd67b99b130781db48e3e1c9e27fd9aea890727c2dc636511547eb9c736: Status 404 returned error can't find the container with id 96fb9fd67b99b130781db48e3e1c9e27fd9aea890727c2dc636511547eb9c736 Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.661602 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-config-data\") pod \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.661660 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-combined-ca-bundle\") pod \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.661963 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-scripts\") pod \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.661992 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/879819c5-1e33-4e31-bdfb-89d74e5efe8b-logs\") pod \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.662071 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks5cl\" (UniqueName: \"kubernetes.io/projected/879819c5-1e33-4e31-bdfb-89d74e5efe8b-kube-api-access-ks5cl\") pod \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\" (UID: \"879819c5-1e33-4e31-bdfb-89d74e5efe8b\") " Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.663381 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/879819c5-1e33-4e31-bdfb-89d74e5efe8b-logs" (OuterVolumeSpecName: "logs") pod "879819c5-1e33-4e31-bdfb-89d74e5efe8b" (UID: "879819c5-1e33-4e31-bdfb-89d74e5efe8b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.667347 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-scripts" (OuterVolumeSpecName: "scripts") pod "879819c5-1e33-4e31-bdfb-89d74e5efe8b" (UID: "879819c5-1e33-4e31-bdfb-89d74e5efe8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.667758 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/879819c5-1e33-4e31-bdfb-89d74e5efe8b-kube-api-access-ks5cl" (OuterVolumeSpecName: "kube-api-access-ks5cl") pod "879819c5-1e33-4e31-bdfb-89d74e5efe8b" (UID: "879819c5-1e33-4e31-bdfb-89d74e5efe8b"). InnerVolumeSpecName "kube-api-access-ks5cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.723893 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-config-data" (OuterVolumeSpecName: "config-data") pod "879819c5-1e33-4e31-bdfb-89d74e5efe8b" (UID: "879819c5-1e33-4e31-bdfb-89d74e5efe8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.723945 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "879819c5-1e33-4e31-bdfb-89d74e5efe8b" (UID: "879819c5-1e33-4e31-bdfb-89d74e5efe8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.764697 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.764740 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/879819c5-1e33-4e31-bdfb-89d74e5efe8b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.764753 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks5cl\" (UniqueName: \"kubernetes.io/projected/879819c5-1e33-4e31-bdfb-89d74e5efe8b-kube-api-access-ks5cl\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.764766 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:03 crc kubenswrapper[4912]: I1203 00:49:03.764777 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/879819c5-1e33-4e31-bdfb-89d74e5efe8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.199886 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59fd659dd5-vmcdc" event={"ID":"6b4100ae-eb0e-402b-b442-d2d313e2bff4","Type":"ContainerStarted","Data":"54885284a185deec5d26e681d75cc88591b9439f6d95d992b94f49c583312784"} Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.200293 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59fd659dd5-vmcdc" event={"ID":"6b4100ae-eb0e-402b-b442-d2d313e2bff4","Type":"ContainerStarted","Data":"96fb9fd67b99b130781db48e3e1c9e27fd9aea890727c2dc636511547eb9c736"} Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.211346 4912 generic.go:334] "Generic (PLEG): container finished" podID="df5c7224-2a57-4931-98b4-78574fb2690b" containerID="ac4118f51a831c59975ddac5999c44968efe58969edd7422dfc30604341e260d" exitCode=0 Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.211596 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x5lrh" event={"ID":"df5c7224-2a57-4931-98b4-78574fb2690b","Type":"ContainerDied","Data":"ac4118f51a831c59975ddac5999c44968efe58969edd7422dfc30604341e260d"} Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.236307 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-wnqwp" event={"ID":"879819c5-1e33-4e31-bdfb-89d74e5efe8b","Type":"ContainerDied","Data":"531b722ca596a5f954ecd07e9d57531836b475038d440ab3cf8d5feda6712883"} Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.236367 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="531b722ca596a5f954ecd07e9d57531836b475038d440ab3cf8d5feda6712883" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.236545 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-wnqwp" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.263759 4912 generic.go:334] "Generic (PLEG): container finished" podID="d4bd0e42-218d-47de-a16f-214d5f4d2910" containerID="10da5eb5c055fa52db79e62a49ca41e7146cd4d94cdcc498e5edc7b6ceb3a6c6" exitCode=0 Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.263847 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dkrd4" event={"ID":"d4bd0e42-218d-47de-a16f-214d5f4d2910","Type":"ContainerDied","Data":"10da5eb5c055fa52db79e62a49ca41e7146cd4d94cdcc498e5edc7b6ceb3a6c6"} Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.301353 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-69d59dff98-dh972"] Dec 03 00:49:04 crc kubenswrapper[4912]: E1203 00:49:04.302081 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="879819c5-1e33-4e31-bdfb-89d74e5efe8b" containerName="placement-db-sync" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.302105 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="879819c5-1e33-4e31-bdfb-89d74e5efe8b" containerName="placement-db-sync" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.302345 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="879819c5-1e33-4e31-bdfb-89d74e5efe8b" containerName="placement-db-sync" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.303778 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.309191 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.309400 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.309459 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.309606 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wcgz7" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.309668 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.320978 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69d59dff98-dh972"] Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.401075 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-logs\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.401147 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-config-data\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.401194 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-public-tls-certs\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.401305 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-scripts\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.401349 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22vkm\" (UniqueName: \"kubernetes.io/projected/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-kube-api-access-22vkm\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.401378 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-combined-ca-bundle\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.401608 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-internal-tls-certs\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.504578 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-internal-tls-certs\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.504663 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-logs\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.504691 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-config-data\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.504725 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-public-tls-certs\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.505315 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-logs\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.506596 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-scripts\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.506749 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22vkm\" (UniqueName: \"kubernetes.io/projected/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-kube-api-access-22vkm\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.506835 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-combined-ca-bundle\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.511598 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-combined-ca-bundle\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.512910 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-internal-tls-certs\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.516935 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-config-data\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.517125 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-public-tls-certs\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.528100 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-scripts\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.539336 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22vkm\" (UniqueName: \"kubernetes.io/projected/7eb76b80-5bef-4a43-a3d0-098c3029fcb2-kube-api-access-22vkm\") pod \"placement-69d59dff98-dh972\" (UID: \"7eb76b80-5bef-4a43-a3d0-098c3029fcb2\") " pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:04 crc kubenswrapper[4912]: I1203 00:49:04.640593 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:06 crc kubenswrapper[4912]: I1203 00:49:06.670862 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:49:06 crc kubenswrapper[4912]: I1203 00:49:06.777484 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4qlxk"] Dec 03 00:49:06 crc kubenswrapper[4912]: I1203 00:49:06.777837 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" podUID="6f2712f0-4d34-4059-b517-51da09b80ce9" containerName="dnsmasq-dns" containerID="cri-o://703d3f45acd79c76aa7eafa8dd2f66e5e31432ddc2f817a46e36cf439ad37dcd" gracePeriod=10 Dec 03 00:49:07 crc kubenswrapper[4912]: I1203 00:49:07.400553 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" podUID="6f2712f0-4d34-4059-b517-51da09b80ce9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.170:5353: connect: connection refused" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.316794 4912 generic.go:334] "Generic (PLEG): container finished" podID="6f2712f0-4d34-4059-b517-51da09b80ce9" containerID="703d3f45acd79c76aa7eafa8dd2f66e5e31432ddc2f817a46e36cf439ad37dcd" exitCode=0 Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.316904 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" event={"ID":"6f2712f0-4d34-4059-b517-51da09b80ce9","Type":"ContainerDied","Data":"703d3f45acd79c76aa7eafa8dd2f66e5e31432ddc2f817a46e36cf439ad37dcd"} Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.321666 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x5lrh" event={"ID":"df5c7224-2a57-4931-98b4-78574fb2690b","Type":"ContainerDied","Data":"cbd07aeb5461c4103bc635d06712bcd30507993223aae3aeb0358ebf52e9acc6"} Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.321735 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbd07aeb5461c4103bc635d06712bcd30507993223aae3aeb0358ebf52e9acc6" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.323401 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dkrd4" event={"ID":"d4bd0e42-218d-47de-a16f-214d5f4d2910","Type":"ContainerDied","Data":"c641c2eb8b20224b86ceb7e079a04b6b4248be8a4134f48cd5fee945cc6093bf"} Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.323531 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c641c2eb8b20224b86ceb7e079a04b6b4248be8a4134f48cd5fee945cc6093bf" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.378444 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.402664 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.438930 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r248j\" (UniqueName: \"kubernetes.io/projected/d4bd0e42-218d-47de-a16f-214d5f4d2910-kube-api-access-r248j\") pod \"d4bd0e42-218d-47de-a16f-214d5f4d2910\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.439347 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-fernet-keys\") pod \"d4bd0e42-218d-47de-a16f-214d5f4d2910\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.439388 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-credential-keys\") pod \"d4bd0e42-218d-47de-a16f-214d5f4d2910\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.439459 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-scripts\") pod \"d4bd0e42-218d-47de-a16f-214d5f4d2910\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.439523 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-config-data\") pod \"d4bd0e42-218d-47de-a16f-214d5f4d2910\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.439547 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-combined-ca-bundle\") pod \"d4bd0e42-218d-47de-a16f-214d5f4d2910\" (UID: \"d4bd0e42-218d-47de-a16f-214d5f4d2910\") " Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.448963 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-scripts" (OuterVolumeSpecName: "scripts") pod "d4bd0e42-218d-47de-a16f-214d5f4d2910" (UID: "d4bd0e42-218d-47de-a16f-214d5f4d2910"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.452284 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4bd0e42-218d-47de-a16f-214d5f4d2910-kube-api-access-r248j" (OuterVolumeSpecName: "kube-api-access-r248j") pod "d4bd0e42-218d-47de-a16f-214d5f4d2910" (UID: "d4bd0e42-218d-47de-a16f-214d5f4d2910"). InnerVolumeSpecName "kube-api-access-r248j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.458362 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d4bd0e42-218d-47de-a16f-214d5f4d2910" (UID: "d4bd0e42-218d-47de-a16f-214d5f4d2910"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.458766 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d4bd0e42-218d-47de-a16f-214d5f4d2910" (UID: "d4bd0e42-218d-47de-a16f-214d5f4d2910"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.506669 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-config-data" (OuterVolumeSpecName: "config-data") pod "d4bd0e42-218d-47de-a16f-214d5f4d2910" (UID: "d4bd0e42-218d-47de-a16f-214d5f4d2910"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.523925 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4bd0e42-218d-47de-a16f-214d5f4d2910" (UID: "d4bd0e42-218d-47de-a16f-214d5f4d2910"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.546105 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-db-sync-config-data\") pod \"df5c7224-2a57-4931-98b4-78574fb2690b\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.546278 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-combined-ca-bundle\") pod \"df5c7224-2a57-4931-98b4-78574fb2690b\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.546349 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwfq4\" (UniqueName: \"kubernetes.io/projected/df5c7224-2a57-4931-98b4-78574fb2690b-kube-api-access-qwfq4\") pod \"df5c7224-2a57-4931-98b4-78574fb2690b\" (UID: \"df5c7224-2a57-4931-98b4-78574fb2690b\") " Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.547928 4912 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.547952 4912 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.547964 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.547976 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.547986 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4bd0e42-218d-47de-a16f-214d5f4d2910-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.547996 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r248j\" (UniqueName: \"kubernetes.io/projected/d4bd0e42-218d-47de-a16f-214d5f4d2910-kube-api-access-r248j\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.551998 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df5c7224-2a57-4931-98b4-78574fb2690b-kube-api-access-qwfq4" (OuterVolumeSpecName: "kube-api-access-qwfq4") pod "df5c7224-2a57-4931-98b4-78574fb2690b" (UID: "df5c7224-2a57-4931-98b4-78574fb2690b"). InnerVolumeSpecName "kube-api-access-qwfq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.555187 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "df5c7224-2a57-4931-98b4-78574fb2690b" (UID: "df5c7224-2a57-4931-98b4-78574fb2690b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.579790 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df5c7224-2a57-4931-98b4-78574fb2690b" (UID: "df5c7224-2a57-4931-98b4-78574fb2690b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.651671 4912 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.651941 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df5c7224-2a57-4931-98b4-78574fb2690b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:08 crc kubenswrapper[4912]: I1203 00:49:08.652000 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwfq4\" (UniqueName: \"kubernetes.io/projected/df5c7224-2a57-4931-98b4-78574fb2690b-kube-api-access-qwfq4\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.333649 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dkrd4" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.333649 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x5lrh" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.497421 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-d9977975b-bq8m5"] Dec 03 00:49:09 crc kubenswrapper[4912]: E1203 00:49:09.498153 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df5c7224-2a57-4931-98b4-78574fb2690b" containerName="barbican-db-sync" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.500494 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="df5c7224-2a57-4931-98b4-78574fb2690b" containerName="barbican-db-sync" Dec 03 00:49:09 crc kubenswrapper[4912]: E1203 00:49:09.500673 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4bd0e42-218d-47de-a16f-214d5f4d2910" containerName="keystone-bootstrap" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.500691 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4bd0e42-218d-47de-a16f-214d5f4d2910" containerName="keystone-bootstrap" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.501266 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4bd0e42-218d-47de-a16f-214d5f4d2910" containerName="keystone-bootstrap" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.501289 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="df5c7224-2a57-4931-98b4-78574fb2690b" containerName="barbican-db-sync" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.502170 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.506097 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.506279 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.506505 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-f2v25" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.506501 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.506624 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.506933 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.534565 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d9977975b-bq8m5"] Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.573129 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-scripts\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.573224 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-config-data\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.573251 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-fernet-keys\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.573305 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nrcc\" (UniqueName: \"kubernetes.io/projected/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-kube-api-access-4nrcc\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.573354 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-credential-keys\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.573376 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-internal-tls-certs\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.573403 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-public-tls-certs\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.573447 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-combined-ca-bundle\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.677953 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-credential-keys\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.678010 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-internal-tls-certs\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.678050 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-public-tls-certs\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.678084 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-combined-ca-bundle\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.678115 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-scripts\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.678247 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-config-data\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.678275 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-fernet-keys\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.678348 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nrcc\" (UniqueName: \"kubernetes.io/projected/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-kube-api-access-4nrcc\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.704418 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-public-tls-certs\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.704463 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-credential-keys\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.704514 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-combined-ca-bundle\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.711201 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-internal-tls-certs\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.721060 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-scripts\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.722735 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-config-data\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.723309 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nrcc\" (UniqueName: \"kubernetes.io/projected/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-kube-api-access-4nrcc\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.731753 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-79459d7cd7-qdq5s"] Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.746274 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b-fernet-keys\") pod \"keystone-d9977975b-bq8m5\" (UID: \"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b\") " pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.748013 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.754107 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2n742" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.754448 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.754588 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.786920 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6f4c446894-wlqxw"] Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.801177 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.805553 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-79459d7cd7-qdq5s"] Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.828086 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.828858 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895327 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-combined-ca-bundle\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895383 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjfbw\" (UniqueName: \"kubernetes.io/projected/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-kube-api-access-bjfbw\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895452 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-config-data\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895501 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-logs\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895536 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-combined-ca-bundle\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895560 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-config-data-custom\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895623 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-logs\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895645 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-config-data-custom\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895708 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8zbj\" (UniqueName: \"kubernetes.io/projected/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-kube-api-access-q8zbj\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.895734 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-config-data\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.953478 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6f4c446894-wlqxw"] Dec 03 00:49:09 crc kubenswrapper[4912]: I1203 00:49:09.985052 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tvc6r"] Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.004405 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.015619 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-combined-ca-bundle\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.015688 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjfbw\" (UniqueName: \"kubernetes.io/projected/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-kube-api-access-bjfbw\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.015729 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-config-data\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.015782 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-logs\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.015819 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-combined-ca-bundle\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.015850 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-config-data-custom\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.015921 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-logs\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.015946 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-config-data-custom\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.016014 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8zbj\" (UniqueName: \"kubernetes.io/projected/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-kube-api-access-q8zbj\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.016039 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-config-data\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.024494 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-logs\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.024512 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-logs\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.031691 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-combined-ca-bundle\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.032390 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-combined-ca-bundle\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.033280 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-config-data\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.035818 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-config-data\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.051911 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-config-data-custom\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.055151 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-config-data-custom\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.060356 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjfbw\" (UniqueName: \"kubernetes.io/projected/3269ec0e-5405-4a8d-b7ea-e07c5a7a7212-kube-api-access-bjfbw\") pod \"barbican-keystone-listener-6f4c446894-wlqxw\" (UID: \"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212\") " pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.072471 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8zbj\" (UniqueName: \"kubernetes.io/projected/c6d0b29e-b4cd-4f78-b9f7-200691c64f93-kube-api-access-q8zbj\") pod \"barbican-worker-79459d7cd7-qdq5s\" (UID: \"c6d0b29e-b4cd-4f78-b9f7-200691c64f93\") " pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.093365 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tvc6r"] Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.118631 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrkxl\" (UniqueName: \"kubernetes.io/projected/2bc22778-372c-43ae-91cb-1007a3163d0d-kube-api-access-vrkxl\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.118742 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.118828 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.118945 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.119154 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.119233 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-config\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.162248 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-79459d7cd7-qdq5s" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.165057 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-95b564956-7sj8f"] Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.167391 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.171052 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.209496 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-95b564956-7sj8f"] Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226214 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226287 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226334 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226371 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-combined-ca-bundle\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226393 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data-custom\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226451 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226473 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226503 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-config\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226586 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6d6m\" (UniqueName: \"kubernetes.io/projected/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-kube-api-access-j6d6m\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226630 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-logs\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.226656 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrkxl\" (UniqueName: \"kubernetes.io/projected/2bc22778-372c-43ae-91cb-1007a3163d0d-kube-api-access-vrkxl\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.228160 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.228933 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.232140 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.234987 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-config\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.237492 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.243054 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.264637 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrkxl\" (UniqueName: \"kubernetes.io/projected/2bc22778-372c-43ae-91cb-1007a3163d0d-kube-api-access-vrkxl\") pod \"dnsmasq-dns-848cf88cfc-tvc6r\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.329036 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-combined-ca-bundle\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.329087 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data-custom\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.329141 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.329207 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6d6m\" (UniqueName: \"kubernetes.io/projected/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-kube-api-access-j6d6m\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.329235 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-logs\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.329941 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-logs\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.338453 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data-custom\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.340584 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.342193 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-combined-ca-bundle\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.362217 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6d6m\" (UniqueName: \"kubernetes.io/projected/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-kube-api-access-j6d6m\") pod \"barbican-api-95b564956-7sj8f\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.393977 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.409067 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.465322 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:10 crc kubenswrapper[4912]: I1203 00:49:10.490938 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.385863 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" event={"ID":"6f2712f0-4d34-4059-b517-51da09b80ce9","Type":"ContainerDied","Data":"9fed1b9772fc502afdc08568599c4f526b9c9ff89e540050cc0cb73df8ad29e4"} Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.386318 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fed1b9772fc502afdc08568599c4f526b9c9ff89e540050cc0cb73df8ad29e4" Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.390860 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.652031 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.762455 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-config\") pod \"6f2712f0-4d34-4059-b517-51da09b80ce9\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.762516 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-swift-storage-0\") pod \"6f2712f0-4d34-4059-b517-51da09b80ce9\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.762616 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97mqn\" (UniqueName: \"kubernetes.io/projected/6f2712f0-4d34-4059-b517-51da09b80ce9-kube-api-access-97mqn\") pod \"6f2712f0-4d34-4059-b517-51da09b80ce9\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.762833 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-svc\") pod \"6f2712f0-4d34-4059-b517-51da09b80ce9\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.762939 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-nb\") pod \"6f2712f0-4d34-4059-b517-51da09b80ce9\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.763015 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-sb\") pod \"6f2712f0-4d34-4059-b517-51da09b80ce9\" (UID: \"6f2712f0-4d34-4059-b517-51da09b80ce9\") " Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.806478 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f2712f0-4d34-4059-b517-51da09b80ce9-kube-api-access-97mqn" (OuterVolumeSpecName: "kube-api-access-97mqn") pod "6f2712f0-4d34-4059-b517-51da09b80ce9" (UID: "6f2712f0-4d34-4059-b517-51da09b80ce9"). InnerVolumeSpecName "kube-api-access-97mqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:11 crc kubenswrapper[4912]: I1203 00:49:11.874583 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97mqn\" (UniqueName: \"kubernetes.io/projected/6f2712f0-4d34-4059-b517-51da09b80ce9-kube-api-access-97mqn\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.027064 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d9977975b-bq8m5"] Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.180414 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6f2712f0-4d34-4059-b517-51da09b80ce9" (UID: "6f2712f0-4d34-4059-b517-51da09b80ce9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.182929 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.248954 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f2712f0-4d34-4059-b517-51da09b80ce9" (UID: "6f2712f0-4d34-4059-b517-51da09b80ce9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.249692 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6f2712f0-4d34-4059-b517-51da09b80ce9" (UID: "6f2712f0-4d34-4059-b517-51da09b80ce9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.291337 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.293450 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.323465 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-config" (OuterVolumeSpecName: "config") pod "6f2712f0-4d34-4059-b517-51da09b80ce9" (UID: "6f2712f0-4d34-4059-b517-51da09b80ce9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.329347 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6f2712f0-4d34-4059-b517-51da09b80ce9" (UID: "6f2712f0-4d34-4059-b517-51da09b80ce9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.398574 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.398631 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f2712f0-4d34-4059-b517-51da09b80ce9-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.417221 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69d59dff98-dh972"] Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.419824 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d9977975b-bq8m5" event={"ID":"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b","Type":"ContainerStarted","Data":"9590014d22dea7ffce028924b263f8d4628494e3b86d1dc3f0de858aa9a65690"} Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.424196 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-4qlxk" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.424517 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59fd659dd5-vmcdc" event={"ID":"6b4100ae-eb0e-402b-b442-d2d313e2bff4","Type":"ContainerStarted","Data":"c375011e2e8f6888754cfeb258c0299e67e0110b2db7ab486f05005ea748b639"} Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.425160 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.438306 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-95b564956-7sj8f"] Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.463201 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-59fd659dd5-vmcdc" podStartSLOduration=10.463177463 podStartE2EDuration="10.463177463s" podCreationTimestamp="2025-12-03 00:49:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:12.455904199 +0000 UTC m=+1538.097924759" watchObservedRunningTime="2025-12-03 00:49:12.463177463 +0000 UTC m=+1538.105198023" Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.669907 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6f4c446894-wlqxw"] Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.694768 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-79459d7cd7-qdq5s"] Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.745422 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4qlxk"] Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.783535 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-4qlxk"] Dec 03 00:49:12 crc kubenswrapper[4912]: I1203 00:49:12.868355 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tvc6r"] Dec 03 00:49:12 crc kubenswrapper[4912]: W1203 00:49:12.869712 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bc22778_372c_43ae_91cb_1007a3163d0d.slice/crio-c90742885165f06b83a62992922662062fd353699d1344ad1b5a8d30445cd432 WatchSource:0}: Error finding container c90742885165f06b83a62992922662062fd353699d1344ad1b5a8d30445cd432: Status 404 returned error can't find the container with id c90742885165f06b83a62992922662062fd353699d1344ad1b5a8d30445cd432 Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.437755 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d9977975b-bq8m5" event={"ID":"a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b","Type":"ContainerStarted","Data":"c832363b0b150d31c3545ecd8de2b586894dd3fdf6824fac779c5fe755c59ea0"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.439895 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.442672 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-95b564956-7sj8f" event={"ID":"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e","Type":"ContainerStarted","Data":"f022956c3e095d4cbedf0a590c2565759de5d9f1252e632d331d2c7aa65ff179"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.442710 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-95b564956-7sj8f" event={"ID":"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e","Type":"ContainerStarted","Data":"08b8ee6f7c345b4e9627df72004e8b0aa5145afd491a51670ee6f0a74a407821"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.449992 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-kwzvj" event={"ID":"bf24aa53-6e61-4aae-88ed-d5085f10a073","Type":"ContainerStarted","Data":"7275f0f547fdd62b6c803cb398bd5a78fa8bd7ce443148ada8771928682cb990"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.465601 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-79459d7cd7-qdq5s" event={"ID":"c6d0b29e-b4cd-4f78-b9f7-200691c64f93","Type":"ContainerStarted","Data":"269bb5de81a409669a47abb55cd3075fd6d7c973c009d3a6ab85893ce39b1722"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.466996 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-d9977975b-bq8m5" podStartSLOduration=4.466979191 podStartE2EDuration="4.466979191s" podCreationTimestamp="2025-12-03 00:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:13.458421582 +0000 UTC m=+1539.100442132" watchObservedRunningTime="2025-12-03 00:49:13.466979191 +0000 UTC m=+1539.108999751" Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.468020 4912 generic.go:334] "Generic (PLEG): container finished" podID="2bc22778-372c-43ae-91cb-1007a3163d0d" containerID="cfb1e0086623df4dc2bbcf49985f4d74c93912cb4b30cdb6dde793d1f2a5c6c9" exitCode=0 Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.470566 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" event={"ID":"2bc22778-372c-43ae-91cb-1007a3163d0d","Type":"ContainerDied","Data":"cfb1e0086623df4dc2bbcf49985f4d74c93912cb4b30cdb6dde793d1f2a5c6c9"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.470606 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" event={"ID":"2bc22778-372c-43ae-91cb-1007a3163d0d","Type":"ContainerStarted","Data":"c90742885165f06b83a62992922662062fd353699d1344ad1b5a8d30445cd432"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.474778 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"616696fb-abad-4f22-949a-b1fa0e87cd90","Type":"ContainerStarted","Data":"f9de75e3c05c329990648266c6e6b659d67ed0ff31d1903c4e0fd280977eda9c"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.486478 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" event={"ID":"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212","Type":"ContainerStarted","Data":"65f8c8097d0fbcbe643d1bdd341a0c6fc9f22a018dd72b41f169df7d5e698adf"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.517683 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-kwzvj" podStartSLOduration=3.823711779 podStartE2EDuration="57.517663254s" podCreationTimestamp="2025-12-03 00:48:16 +0000 UTC" firstStartedPulling="2025-12-03 00:48:18.120684228 +0000 UTC m=+1483.762704788" lastFinishedPulling="2025-12-03 00:49:11.814635703 +0000 UTC m=+1537.456656263" observedRunningTime="2025-12-03 00:49:13.487227872 +0000 UTC m=+1539.129248452" watchObservedRunningTime="2025-12-03 00:49:13.517663254 +0000 UTC m=+1539.159683814" Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.518318 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69d59dff98-dh972" event={"ID":"7eb76b80-5bef-4a43-a3d0-098c3029fcb2","Type":"ContainerStarted","Data":"29e8dc79a22964759cd30515c86a15971770870838c823aa5c80c1a6c2516ac0"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.518580 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69d59dff98-dh972" event={"ID":"7eb76b80-5bef-4a43-a3d0-098c3029fcb2","Type":"ContainerStarted","Data":"919e2f5cbc2309e807ee97ffbef28522ea86ac024100e162979de62f6362f45a"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.518598 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69d59dff98-dh972" event={"ID":"7eb76b80-5bef-4a43-a3d0-098c3029fcb2","Type":"ContainerStarted","Data":"d3dc09baa731c825b15e3472edf4404c7b3e16869c98ef1b288a7ca0e4dc8d35"} Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.518682 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.519671 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:13 crc kubenswrapper[4912]: I1203 00:49:13.661896 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-69d59dff98-dh972" podStartSLOduration=9.661864795 podStartE2EDuration="9.661864795s" podCreationTimestamp="2025-12-03 00:49:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:13.554486168 +0000 UTC m=+1539.196506728" watchObservedRunningTime="2025-12-03 00:49:13.661864795 +0000 UTC m=+1539.303885365" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.043622 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-75f549cc78-m5rcp"] Dec 03 00:49:14 crc kubenswrapper[4912]: E1203 00:49:14.044194 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f2712f0-4d34-4059-b517-51da09b80ce9" containerName="init" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.044216 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f2712f0-4d34-4059-b517-51da09b80ce9" containerName="init" Dec 03 00:49:14 crc kubenswrapper[4912]: E1203 00:49:14.044241 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f2712f0-4d34-4059-b517-51da09b80ce9" containerName="dnsmasq-dns" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.044247 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f2712f0-4d34-4059-b517-51da09b80ce9" containerName="dnsmasq-dns" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.051616 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f2712f0-4d34-4059-b517-51da09b80ce9" containerName="dnsmasq-dns" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.053114 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.067402 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.067719 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.113472 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75f549cc78-m5rcp"] Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.190243 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-combined-ca-bundle\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.190349 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e2b86e-eeab-4af6-b68d-8f67abda11e2-logs\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.190405 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-public-tls-certs\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.190448 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-internal-tls-certs\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.190488 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-config-data\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.190519 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-config-data-custom\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.190589 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qkd5\" (UniqueName: \"kubernetes.io/projected/21e2b86e-eeab-4af6-b68d-8f67abda11e2-kube-api-access-7qkd5\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.293614 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-combined-ca-bundle\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.293704 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e2b86e-eeab-4af6-b68d-8f67abda11e2-logs\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.293761 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-public-tls-certs\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.293781 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-internal-tls-certs\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.293825 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-config-data\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.294049 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-config-data-custom\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.294127 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qkd5\" (UniqueName: \"kubernetes.io/projected/21e2b86e-eeab-4af6-b68d-8f67abda11e2-kube-api-access-7qkd5\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.295378 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e2b86e-eeab-4af6-b68d-8f67abda11e2-logs\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.307032 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-combined-ca-bundle\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.318395 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-config-data-custom\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.320034 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-config-data\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.321583 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-public-tls-certs\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.323004 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e2b86e-eeab-4af6-b68d-8f67abda11e2-internal-tls-certs\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.331155 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qkd5\" (UniqueName: \"kubernetes.io/projected/21e2b86e-eeab-4af6-b68d-8f67abda11e2-kube-api-access-7qkd5\") pod \"barbican-api-75f549cc78-m5rcp\" (UID: \"21e2b86e-eeab-4af6-b68d-8f67abda11e2\") " pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.430475 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.543085 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-b5xlq" event={"ID":"3071752c-9584-4ac1-8746-1b3a4ad3f4a1","Type":"ContainerStarted","Data":"94a4977e9785851db0df0bd8592b4ba0d031198bade66107dccab490141f0daf"} Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.544888 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-95b564956-7sj8f" event={"ID":"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e","Type":"ContainerStarted","Data":"b199cb1e084a4ef25b9e2754a30d6236eeb6172edd6f45aae0e63b28d65866f7"} Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.546539 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.546611 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.560799 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" event={"ID":"2bc22778-372c-43ae-91cb-1007a3163d0d","Type":"ContainerStarted","Data":"03ebca25b12667bf34b1085e1c7e3d2b91419437061ea1f4d17c9f80a2e3975c"} Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.560887 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.578603 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-95b564956-7sj8f" podStartSLOduration=4.5785719369999995 podStartE2EDuration="4.578571937s" podCreationTimestamp="2025-12-03 00:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:14.576803411 +0000 UTC m=+1540.218823981" watchObservedRunningTime="2025-12-03 00:49:14.578571937 +0000 UTC m=+1540.220592497" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.602650 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f2712f0-4d34-4059-b517-51da09b80ce9" path="/var/lib/kubelet/pods/6f2712f0-4d34-4059-b517-51da09b80ce9/volumes" Dec 03 00:49:14 crc kubenswrapper[4912]: I1203 00:49:14.625701 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" podStartSLOduration=5.625660545 podStartE2EDuration="5.625660545s" podCreationTimestamp="2025-12-03 00:49:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:14.605142187 +0000 UTC m=+1540.247162767" watchObservedRunningTime="2025-12-03 00:49:14.625660545 +0000 UTC m=+1540.267681105" Dec 03 00:49:15 crc kubenswrapper[4912]: I1203 00:49:15.000132 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75f549cc78-m5rcp"] Dec 03 00:49:15 crc kubenswrapper[4912]: W1203 00:49:15.010609 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21e2b86e_eeab_4af6_b68d_8f67abda11e2.slice/crio-3c526b0698866d9908daa9ff25a0a16debb0107bb1ace85fe592ac297722ef4b WatchSource:0}: Error finding container 3c526b0698866d9908daa9ff25a0a16debb0107bb1ace85fe592ac297722ef4b: Status 404 returned error can't find the container with id 3c526b0698866d9908daa9ff25a0a16debb0107bb1ace85fe592ac297722ef4b Dec 03 00:49:15 crc kubenswrapper[4912]: I1203 00:49:15.572795 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75f549cc78-m5rcp" event={"ID":"21e2b86e-eeab-4af6-b68d-8f67abda11e2","Type":"ContainerStarted","Data":"279019e6f1d02768ce0d2936192386717b4849ed602b915b5ff7569fd35dfe95"} Dec 03 00:49:15 crc kubenswrapper[4912]: I1203 00:49:15.573693 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75f549cc78-m5rcp" event={"ID":"21e2b86e-eeab-4af6-b68d-8f67abda11e2","Type":"ContainerStarted","Data":"3c526b0698866d9908daa9ff25a0a16debb0107bb1ace85fe592ac297722ef4b"} Dec 03 00:49:15 crc kubenswrapper[4912]: I1203 00:49:15.607828 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-b5xlq" podStartSLOduration=5.399505295 podStartE2EDuration="59.607801195s" podCreationTimestamp="2025-12-03 00:48:16 +0000 UTC" firstStartedPulling="2025-12-03 00:48:18.951798015 +0000 UTC m=+1484.593818575" lastFinishedPulling="2025-12-03 00:49:13.160093915 +0000 UTC m=+1538.802114475" observedRunningTime="2025-12-03 00:49:15.60502429 +0000 UTC m=+1541.247044860" watchObservedRunningTime="2025-12-03 00:49:15.607801195 +0000 UTC m=+1541.249821755" Dec 03 00:49:17 crc kubenswrapper[4912]: I1203 00:49:17.673548 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75f549cc78-m5rcp" event={"ID":"21e2b86e-eeab-4af6-b68d-8f67abda11e2","Type":"ContainerStarted","Data":"d5b577626ecb1a7a38cee694bff7c7814173502d903198a57dfc66a6ee2d36cd"} Dec 03 00:49:17 crc kubenswrapper[4912]: I1203 00:49:17.674221 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:17 crc kubenswrapper[4912]: I1203 00:49:17.677380 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" event={"ID":"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212","Type":"ContainerStarted","Data":"47c12d49d8d9434d35d4bd2a035387ca37264e5fd07690ea2ddac0ca22239980"} Dec 03 00:49:17 crc kubenswrapper[4912]: I1203 00:49:17.677423 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" event={"ID":"3269ec0e-5405-4a8d-b7ea-e07c5a7a7212","Type":"ContainerStarted","Data":"3ce8a1b5ecb2ca6a5f1444640b86da4d178bf14014d1684c5ba4706ef85b95ce"} Dec 03 00:49:17 crc kubenswrapper[4912]: I1203 00:49:17.680931 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-79459d7cd7-qdq5s" event={"ID":"c6d0b29e-b4cd-4f78-b9f7-200691c64f93","Type":"ContainerStarted","Data":"580481c84c3b7d5a49e01fae803a4ae0fac02e04a7185f972cc2bef93bbe2f0e"} Dec 03 00:49:17 crc kubenswrapper[4912]: I1203 00:49:17.680972 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-79459d7cd7-qdq5s" event={"ID":"c6d0b29e-b4cd-4f78-b9f7-200691c64f93","Type":"ContainerStarted","Data":"160fc5852d1364a93d9b724968c53421f733d72fb7607c109456f42c289a5272"} Dec 03 00:49:17 crc kubenswrapper[4912]: I1203 00:49:17.713346 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-75f549cc78-m5rcp" podStartSLOduration=3.713300095 podStartE2EDuration="3.713300095s" podCreationTimestamp="2025-12-03 00:49:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:17.693765143 +0000 UTC m=+1543.335785713" watchObservedRunningTime="2025-12-03 00:49:17.713300095 +0000 UTC m=+1543.355320655" Dec 03 00:49:17 crc kubenswrapper[4912]: I1203 00:49:17.742722 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-79459d7cd7-qdq5s" podStartSLOduration=4.893050229 podStartE2EDuration="8.74270103s" podCreationTimestamp="2025-12-03 00:49:09 +0000 UTC" firstStartedPulling="2025-12-03 00:49:12.802486824 +0000 UTC m=+1538.444507384" lastFinishedPulling="2025-12-03 00:49:16.652137625 +0000 UTC m=+1542.294158185" observedRunningTime="2025-12-03 00:49:17.718923575 +0000 UTC m=+1543.360944145" watchObservedRunningTime="2025-12-03 00:49:17.74270103 +0000 UTC m=+1543.384721590" Dec 03 00:49:17 crc kubenswrapper[4912]: I1203 00:49:17.758966 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6f4c446894-wlqxw" podStartSLOduration=4.908747727 podStartE2EDuration="8.758944403s" podCreationTimestamp="2025-12-03 00:49:09 +0000 UTC" firstStartedPulling="2025-12-03 00:49:12.800691046 +0000 UTC m=+1538.442711606" lastFinishedPulling="2025-12-03 00:49:16.650887722 +0000 UTC m=+1542.292908282" observedRunningTime="2025-12-03 00:49:17.741051236 +0000 UTC m=+1543.383071806" watchObservedRunningTime="2025-12-03 00:49:17.758944403 +0000 UTC m=+1543.400964963" Dec 03 00:49:18 crc kubenswrapper[4912]: I1203 00:49:18.077862 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:49:18 crc kubenswrapper[4912]: I1203 00:49:18.077927 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:49:18 crc kubenswrapper[4912]: I1203 00:49:18.701660 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:20 crc kubenswrapper[4912]: I1203 00:49:20.468360 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:20 crc kubenswrapper[4912]: I1203 00:49:20.564777 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6b98f"] Dec 03 00:49:20 crc kubenswrapper[4912]: I1203 00:49:20.565124 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" podUID="dc3a3200-2b91-41bb-8a80-25a8d2601331" containerName="dnsmasq-dns" containerID="cri-o://c02d91c0179dfcf5c01d4d143c4d8a4bf3c0327e0171a843603c1402cdd86248" gracePeriod=10 Dec 03 00:49:20 crc kubenswrapper[4912]: I1203 00:49:20.735665 4912 generic.go:334] "Generic (PLEG): container finished" podID="bf24aa53-6e61-4aae-88ed-d5085f10a073" containerID="7275f0f547fdd62b6c803cb398bd5a78fa8bd7ce443148ada8771928682cb990" exitCode=0 Dec 03 00:49:20 crc kubenswrapper[4912]: I1203 00:49:20.735711 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-kwzvj" event={"ID":"bf24aa53-6e61-4aae-88ed-d5085f10a073","Type":"ContainerDied","Data":"7275f0f547fdd62b6c803cb398bd5a78fa8bd7ce443148ada8771928682cb990"} Dec 03 00:49:21 crc kubenswrapper[4912]: I1203 00:49:21.669579 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" podUID="dc3a3200-2b91-41bb-8a80-25a8d2601331" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: connect: connection refused" Dec 03 00:49:21 crc kubenswrapper[4912]: I1203 00:49:21.755020 4912 generic.go:334] "Generic (PLEG): container finished" podID="dc3a3200-2b91-41bb-8a80-25a8d2601331" containerID="c02d91c0179dfcf5c01d4d143c4d8a4bf3c0327e0171a843603c1402cdd86248" exitCode=0 Dec 03 00:49:21 crc kubenswrapper[4912]: I1203 00:49:21.755135 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" event={"ID":"dc3a3200-2b91-41bb-8a80-25a8d2601331","Type":"ContainerDied","Data":"c02d91c0179dfcf5c01d4d143c4d8a4bf3c0327e0171a843603c1402cdd86248"} Dec 03 00:49:22 crc kubenswrapper[4912]: I1203 00:49:22.273216 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:22 crc kubenswrapper[4912]: I1203 00:49:22.308890 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:22 crc kubenswrapper[4912]: I1203 00:49:22.773240 4912 generic.go:334] "Generic (PLEG): container finished" podID="3071752c-9584-4ac1-8746-1b3a4ad3f4a1" containerID="94a4977e9785851db0df0bd8592b4ba0d031198bade66107dccab490141f0daf" exitCode=0 Dec 03 00:49:22 crc kubenswrapper[4912]: I1203 00:49:22.773293 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-b5xlq" event={"ID":"3071752c-9584-4ac1-8746-1b3a4ad3f4a1","Type":"ContainerDied","Data":"94a4977e9785851db0df0bd8592b4ba0d031198bade66107dccab490141f0daf"} Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.260002 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-kwzvj" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.286328 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:49:23 crc kubenswrapper[4912]: E1203 00:49:23.380374 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.385537 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-config\") pod \"dc3a3200-2b91-41bb-8a80-25a8d2601331\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.385606 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-combined-ca-bundle\") pod \"bf24aa53-6e61-4aae-88ed-d5085f10a073\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.385692 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-config-data\") pod \"bf24aa53-6e61-4aae-88ed-d5085f10a073\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.385737 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-svc\") pod \"dc3a3200-2b91-41bb-8a80-25a8d2601331\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.385760 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f24zm\" (UniqueName: \"kubernetes.io/projected/dc3a3200-2b91-41bb-8a80-25a8d2601331-kube-api-access-f24zm\") pod \"dc3a3200-2b91-41bb-8a80-25a8d2601331\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.385833 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldqpn\" (UniqueName: \"kubernetes.io/projected/bf24aa53-6e61-4aae-88ed-d5085f10a073-kube-api-access-ldqpn\") pod \"bf24aa53-6e61-4aae-88ed-d5085f10a073\" (UID: \"bf24aa53-6e61-4aae-88ed-d5085f10a073\") " Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.385882 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-swift-storage-0\") pod \"dc3a3200-2b91-41bb-8a80-25a8d2601331\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.385921 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-sb\") pod \"dc3a3200-2b91-41bb-8a80-25a8d2601331\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.386004 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-nb\") pod \"dc3a3200-2b91-41bb-8a80-25a8d2601331\" (UID: \"dc3a3200-2b91-41bb-8a80-25a8d2601331\") " Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.398755 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf24aa53-6e61-4aae-88ed-d5085f10a073-kube-api-access-ldqpn" (OuterVolumeSpecName: "kube-api-access-ldqpn") pod "bf24aa53-6e61-4aae-88ed-d5085f10a073" (UID: "bf24aa53-6e61-4aae-88ed-d5085f10a073"). InnerVolumeSpecName "kube-api-access-ldqpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.404072 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc3a3200-2b91-41bb-8a80-25a8d2601331-kube-api-access-f24zm" (OuterVolumeSpecName: "kube-api-access-f24zm") pod "dc3a3200-2b91-41bb-8a80-25a8d2601331" (UID: "dc3a3200-2b91-41bb-8a80-25a8d2601331"). InnerVolumeSpecName "kube-api-access-f24zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.433560 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf24aa53-6e61-4aae-88ed-d5085f10a073" (UID: "bf24aa53-6e61-4aae-88ed-d5085f10a073"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.469392 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dc3a3200-2b91-41bb-8a80-25a8d2601331" (UID: "dc3a3200-2b91-41bb-8a80-25a8d2601331"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.476010 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-config" (OuterVolumeSpecName: "config") pod "dc3a3200-2b91-41bb-8a80-25a8d2601331" (UID: "dc3a3200-2b91-41bb-8a80-25a8d2601331"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.483046 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dc3a3200-2b91-41bb-8a80-25a8d2601331" (UID: "dc3a3200-2b91-41bb-8a80-25a8d2601331"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.486575 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dc3a3200-2b91-41bb-8a80-25a8d2601331" (UID: "dc3a3200-2b91-41bb-8a80-25a8d2601331"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.493111 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldqpn\" (UniqueName: \"kubernetes.io/projected/bf24aa53-6e61-4aae-88ed-d5085f10a073-kube-api-access-ldqpn\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.493134 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.493145 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.493158 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.493170 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.493179 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.493187 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f24zm\" (UniqueName: \"kubernetes.io/projected/dc3a3200-2b91-41bb-8a80-25a8d2601331-kube-api-access-f24zm\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.494183 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dc3a3200-2b91-41bb-8a80-25a8d2601331" (UID: "dc3a3200-2b91-41bb-8a80-25a8d2601331"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.524645 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-config-data" (OuterVolumeSpecName: "config-data") pod "bf24aa53-6e61-4aae-88ed-d5085f10a073" (UID: "bf24aa53-6e61-4aae-88ed-d5085f10a073"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.595510 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dc3a3200-2b91-41bb-8a80-25a8d2601331-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.595558 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf24aa53-6e61-4aae-88ed-d5085f10a073-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.792190 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" event={"ID":"dc3a3200-2b91-41bb-8a80-25a8d2601331","Type":"ContainerDied","Data":"042c550ebe94a08a8485d5caeea879862fac6b6ef1d5325b9f6c7888c97516b8"} Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.792259 4912 scope.go:117] "RemoveContainer" containerID="c02d91c0179dfcf5c01d4d143c4d8a4bf3c0327e0171a843603c1402cdd86248" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.792414 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-6b98f" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.795859 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-kwzvj" event={"ID":"bf24aa53-6e61-4aae-88ed-d5085f10a073","Type":"ContainerDied","Data":"58de2bb02e247baabc8ff591931e1a1fafa129ba778b987f88c906d31603b1c0"} Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.795922 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58de2bb02e247baabc8ff591931e1a1fafa129ba778b987f88c906d31603b1c0" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.795924 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-kwzvj" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.799491 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"616696fb-abad-4f22-949a-b1fa0e87cd90","Type":"ContainerStarted","Data":"863faecd86ce70d02cb5598f0cac5a5c04f195b242803006c751c1daef68c275"} Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.799728 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="ceilometer-notification-agent" containerID="cri-o://2a585e40cc6f95d08d378351d5a1ca2a190d4059b133042e0dbe2586ddfef79b" gracePeriod=30 Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.799743 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="proxy-httpd" containerID="cri-o://863faecd86ce70d02cb5598f0cac5a5c04f195b242803006c751c1daef68c275" gracePeriod=30 Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.799806 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.799804 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="sg-core" containerID="cri-o://f9de75e3c05c329990648266c6e6b659d67ed0ff31d1903c4e0fd280977eda9c" gracePeriod=30 Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.859676 4912 scope.go:117] "RemoveContainer" containerID="4a157c5e14db42e4d5f3cb81957deeeb118b425f226e6621ee24676bed2a85d2" Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.881275 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6b98f"] Dec 03 00:49:23 crc kubenswrapper[4912]: I1203 00:49:23.894489 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-6b98f"] Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.191783 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.324392 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-scripts\") pod \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.324629 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd4gm\" (UniqueName: \"kubernetes.io/projected/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-kube-api-access-qd4gm\") pod \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.324781 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-config-data\") pod \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.324819 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-db-sync-config-data\") pod \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.324864 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-combined-ca-bundle\") pod \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.324959 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-etc-machine-id\") pod \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\" (UID: \"3071752c-9584-4ac1-8746-1b3a4ad3f4a1\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.325302 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3071752c-9584-4ac1-8746-1b3a4ad3f4a1" (UID: "3071752c-9584-4ac1-8746-1b3a4ad3f4a1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.325659 4912 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.333261 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3071752c-9584-4ac1-8746-1b3a4ad3f4a1" (UID: "3071752c-9584-4ac1-8746-1b3a4ad3f4a1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.333767 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-scripts" (OuterVolumeSpecName: "scripts") pod "3071752c-9584-4ac1-8746-1b3a4ad3f4a1" (UID: "3071752c-9584-4ac1-8746-1b3a4ad3f4a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.336718 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-kube-api-access-qd4gm" (OuterVolumeSpecName: "kube-api-access-qd4gm") pod "3071752c-9584-4ac1-8746-1b3a4ad3f4a1" (UID: "3071752c-9584-4ac1-8746-1b3a4ad3f4a1"). InnerVolumeSpecName "kube-api-access-qd4gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.363043 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3071752c-9584-4ac1-8746-1b3a4ad3f4a1" (UID: "3071752c-9584-4ac1-8746-1b3a4ad3f4a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.404018 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-config-data" (OuterVolumeSpecName: "config-data") pod "3071752c-9584-4ac1-8746-1b3a4ad3f4a1" (UID: "3071752c-9584-4ac1-8746-1b3a4ad3f4a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.427924 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.427956 4912 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.427966 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.427975 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.427984 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd4gm\" (UniqueName: \"kubernetes.io/projected/3071752c-9584-4ac1-8746-1b3a4ad3f4a1-kube-api-access-qd4gm\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.663639 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc3a3200-2b91-41bb-8a80-25a8d2601331" path="/var/lib/kubelet/pods/dc3a3200-2b91-41bb-8a80-25a8d2601331/volumes" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.816369 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-b5xlq" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.816500 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-b5xlq" event={"ID":"3071752c-9584-4ac1-8746-1b3a4ad3f4a1","Type":"ContainerDied","Data":"eaf4e20ba560bcc1cbb1f471495ed9ffa600a1d6b4be322cfdcfb3bf73948294"} Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.816550 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaf4e20ba560bcc1cbb1f471495ed9ffa600a1d6b4be322cfdcfb3bf73948294" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.823468 4912 generic.go:334] "Generic (PLEG): container finished" podID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerID="863faecd86ce70d02cb5598f0cac5a5c04f195b242803006c751c1daef68c275" exitCode=0 Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.823504 4912 generic.go:334] "Generic (PLEG): container finished" podID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerID="f9de75e3c05c329990648266c6e6b659d67ed0ff31d1903c4e0fd280977eda9c" exitCode=2 Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.823512 4912 generic.go:334] "Generic (PLEG): container finished" podID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerID="2a585e40cc6f95d08d378351d5a1ca2a190d4059b133042e0dbe2586ddfef79b" exitCode=0 Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.823535 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"616696fb-abad-4f22-949a-b1fa0e87cd90","Type":"ContainerDied","Data":"863faecd86ce70d02cb5598f0cac5a5c04f195b242803006c751c1daef68c275"} Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.823566 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"616696fb-abad-4f22-949a-b1fa0e87cd90","Type":"ContainerDied","Data":"f9de75e3c05c329990648266c6e6b659d67ed0ff31d1903c4e0fd280977eda9c"} Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.823576 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"616696fb-abad-4f22-949a-b1fa0e87cd90","Type":"ContainerDied","Data":"2a585e40cc6f95d08d378351d5a1ca2a190d4059b133042e0dbe2586ddfef79b"} Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.823584 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"616696fb-abad-4f22-949a-b1fa0e87cd90","Type":"ContainerDied","Data":"9f9ad0422fd275be64df81cb085ddfc7fc57f707f63af227b4b0288994fc5a9f"} Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.823594 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f9ad0422fd275be64df81cb085ddfc7fc57f707f63af227b4b0288994fc5a9f" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.824175 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.942018 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-run-httpd\") pod \"616696fb-abad-4f22-949a-b1fa0e87cd90\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.942123 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4rrc\" (UniqueName: \"kubernetes.io/projected/616696fb-abad-4f22-949a-b1fa0e87cd90-kube-api-access-p4rrc\") pod \"616696fb-abad-4f22-949a-b1fa0e87cd90\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.942156 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-scripts\") pod \"616696fb-abad-4f22-949a-b1fa0e87cd90\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.942230 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-log-httpd\") pod \"616696fb-abad-4f22-949a-b1fa0e87cd90\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.942313 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-sg-core-conf-yaml\") pod \"616696fb-abad-4f22-949a-b1fa0e87cd90\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.942611 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-combined-ca-bundle\") pod \"616696fb-abad-4f22-949a-b1fa0e87cd90\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.942625 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "616696fb-abad-4f22-949a-b1fa0e87cd90" (UID: "616696fb-abad-4f22-949a-b1fa0e87cd90"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.942712 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-config-data\") pod \"616696fb-abad-4f22-949a-b1fa0e87cd90\" (UID: \"616696fb-abad-4f22-949a-b1fa0e87cd90\") " Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.942814 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "616696fb-abad-4f22-949a-b1fa0e87cd90" (UID: "616696fb-abad-4f22-949a-b1fa0e87cd90"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.943932 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.943962 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/616696fb-abad-4f22-949a-b1fa0e87cd90-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.947906 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-scripts" (OuterVolumeSpecName: "scripts") pod "616696fb-abad-4f22-949a-b1fa0e87cd90" (UID: "616696fb-abad-4f22-949a-b1fa0e87cd90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:24 crc kubenswrapper[4912]: I1203 00:49:24.958770 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/616696fb-abad-4f22-949a-b1fa0e87cd90-kube-api-access-p4rrc" (OuterVolumeSpecName: "kube-api-access-p4rrc") pod "616696fb-abad-4f22-949a-b1fa0e87cd90" (UID: "616696fb-abad-4f22-949a-b1fa0e87cd90"). InnerVolumeSpecName "kube-api-access-p4rrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.052861 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "616696fb-abad-4f22-949a-b1fa0e87cd90" (UID: "616696fb-abad-4f22-949a-b1fa0e87cd90"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.056098 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4rrc\" (UniqueName: \"kubernetes.io/projected/616696fb-abad-4f22-949a-b1fa0e87cd90-kube-api-access-p4rrc\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.056153 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.056168 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.116571 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 00:49:25 crc kubenswrapper[4912]: E1203 00:49:25.117196 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="proxy-httpd" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117213 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="proxy-httpd" Dec 03 00:49:25 crc kubenswrapper[4912]: E1203 00:49:25.117229 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf24aa53-6e61-4aae-88ed-d5085f10a073" containerName="heat-db-sync" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117237 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf24aa53-6e61-4aae-88ed-d5085f10a073" containerName="heat-db-sync" Dec 03 00:49:25 crc kubenswrapper[4912]: E1203 00:49:25.117289 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="sg-core" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117297 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="sg-core" Dec 03 00:49:25 crc kubenswrapper[4912]: E1203 00:49:25.117306 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3071752c-9584-4ac1-8746-1b3a4ad3f4a1" containerName="cinder-db-sync" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117314 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="3071752c-9584-4ac1-8746-1b3a4ad3f4a1" containerName="cinder-db-sync" Dec 03 00:49:25 crc kubenswrapper[4912]: E1203 00:49:25.117327 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3a3200-2b91-41bb-8a80-25a8d2601331" containerName="dnsmasq-dns" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117335 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3a3200-2b91-41bb-8a80-25a8d2601331" containerName="dnsmasq-dns" Dec 03 00:49:25 crc kubenswrapper[4912]: E1203 00:49:25.117345 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="ceilometer-notification-agent" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117351 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="ceilometer-notification-agent" Dec 03 00:49:25 crc kubenswrapper[4912]: E1203 00:49:25.117362 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc3a3200-2b91-41bb-8a80-25a8d2601331" containerName="init" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117370 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc3a3200-2b91-41bb-8a80-25a8d2601331" containerName="init" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117583 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="ceilometer-notification-agent" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117596 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="proxy-httpd" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117604 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc3a3200-2b91-41bb-8a80-25a8d2601331" containerName="dnsmasq-dns" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117623 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf24aa53-6e61-4aae-88ed-d5085f10a073" containerName="heat-db-sync" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117631 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="3071752c-9584-4ac1-8746-1b3a4ad3f4a1" containerName="cinder-db-sync" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.117637 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" containerName="sg-core" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.129539 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.143109 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.143358 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.143474 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-frht4" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.143570 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.144541 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-config-data" (OuterVolumeSpecName: "config-data") pod "616696fb-abad-4f22-949a-b1fa0e87cd90" (UID: "616696fb-abad-4f22-949a-b1fa0e87cd90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.158614 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.183350 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.189791 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "616696fb-abad-4f22-949a-b1fa0e87cd90" (UID: "616696fb-abad-4f22-949a-b1fa0e87cd90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.206975 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-zmrhq"] Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.209424 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.259357 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-zmrhq"] Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.261891 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.261926 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.261990 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-config\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262028 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262055 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-scripts\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262132 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262155 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfpmr\" (UniqueName: \"kubernetes.io/projected/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-kube-api-access-nfpmr\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262178 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz44b\" (UniqueName: \"kubernetes.io/projected/26f3f384-6499-4824-8c10-de6382e27d8b-kube-api-access-xz44b\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262199 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262215 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262255 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262282 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.262345 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/616696fb-abad-4f22-949a-b1fa0e87cd90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.325865 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.334009 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.336661 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.337685 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.364671 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.364720 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.364757 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22763c05-583a-4681-bbf4-882fd2a8d7ed-etc-machine-id\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.364802 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22763c05-583a-4681-bbf4-882fd2a8d7ed-logs\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.364832 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-config\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.364869 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4488v\" (UniqueName: \"kubernetes.io/projected/22763c05-583a-4681-bbf4-882fd2a8d7ed-kube-api-access-4488v\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.364886 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.364905 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.364932 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-scripts\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.365006 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data-custom\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.365033 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.365054 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.365076 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfpmr\" (UniqueName: \"kubernetes.io/projected/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-kube-api-access-nfpmr\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.365098 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz44b\" (UniqueName: \"kubernetes.io/projected/26f3f384-6499-4824-8c10-de6382e27d8b-kube-api-access-xz44b\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.365121 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.365140 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.365180 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.366983 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.367211 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-svc\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.367667 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-scripts\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.367750 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.368068 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.368715 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.368826 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-config\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.369833 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.375691 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.379167 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.383626 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.388239 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-scripts\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.391244 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz44b\" (UniqueName: \"kubernetes.io/projected/26f3f384-6499-4824-8c10-de6382e27d8b-kube-api-access-xz44b\") pod \"dnsmasq-dns-6578955fd5-zmrhq\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.397128 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfpmr\" (UniqueName: \"kubernetes.io/projected/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-kube-api-access-nfpmr\") pod \"cinder-scheduler-0\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.469768 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4488v\" (UniqueName: \"kubernetes.io/projected/22763c05-583a-4681-bbf4-882fd2a8d7ed-kube-api-access-4488v\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.469833 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.469924 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data-custom\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.469952 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.470022 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-scripts\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.470070 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22763c05-583a-4681-bbf4-882fd2a8d7ed-etc-machine-id\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.470110 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22763c05-583a-4681-bbf4-882fd2a8d7ed-logs\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.470609 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22763c05-583a-4681-bbf4-882fd2a8d7ed-logs\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.470965 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22763c05-583a-4681-bbf4-882fd2a8d7ed-etc-machine-id\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.474723 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.475249 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data-custom\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.477570 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-scripts\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.477969 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.490457 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4488v\" (UniqueName: \"kubernetes.io/projected/22763c05-583a-4681-bbf4-882fd2a8d7ed-kube-api-access-4488v\") pod \"cinder-api-0\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.516795 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.553411 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.654217 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.839967 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:49:25 crc kubenswrapper[4912]: I1203 00:49:25.973606 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.014496 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.039943 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.045571 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.050333 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.050376 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.088122 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.100597 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-scripts\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.100791 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-run-httpd\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.100860 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.100880 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.100940 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46p4d\" (UniqueName: \"kubernetes.io/projected/caa68042-1653-4df5-9249-6263d3fcbd52-kube-api-access-46p4d\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.100991 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-log-httpd\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.101043 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-config-data\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: W1203 00:49:26.105560 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaffbee0c_a6f2_4e2c_a7fe_c9fc2bd446fa.slice/crio-c94935886c4e8dce7700d534d888ba822644d7b859b6b66d8deb6c8cc3470d0f WatchSource:0}: Error finding container c94935886c4e8dce7700d534d888ba822644d7b859b6b66d8deb6c8cc3470d0f: Status 404 returned error can't find the container with id c94935886c4e8dce7700d534d888ba822644d7b859b6b66d8deb6c8cc3470d0f Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.141335 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-zmrhq"] Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.163408 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.206599 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.206646 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.206717 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46p4d\" (UniqueName: \"kubernetes.io/projected/caa68042-1653-4df5-9249-6263d3fcbd52-kube-api-access-46p4d\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.206766 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-log-httpd\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.206811 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-config-data\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.206894 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-scripts\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.206993 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-run-httpd\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.207673 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-run-httpd\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.209408 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-log-httpd\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.222572 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.223880 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-config-data\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.224879 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-scripts\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.230904 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.238859 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46p4d\" (UniqueName: \"kubernetes.io/projected/caa68042-1653-4df5-9249-6263d3fcbd52-kube-api-access-46p4d\") pod \"ceilometer-0\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.405214 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.416872 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.564341 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.602937 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="616696fb-abad-4f22-949a-b1fa0e87cd90" path="/var/lib/kubelet/pods/616696fb-abad-4f22-949a-b1fa0e87cd90/volumes" Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.904824 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa","Type":"ContainerStarted","Data":"c94935886c4e8dce7700d534d888ba822644d7b859b6b66d8deb6c8cc3470d0f"} Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.930822 4912 generic.go:334] "Generic (PLEG): container finished" podID="26f3f384-6499-4824-8c10-de6382e27d8b" containerID="3a505aa2448fb487f9c9d0839f071622b20dd2151dcb926fb2dd7ccf53da3a9e" exitCode=0 Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.931370 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" event={"ID":"26f3f384-6499-4824-8c10-de6382e27d8b","Type":"ContainerDied","Data":"3a505aa2448fb487f9c9d0839f071622b20dd2151dcb926fb2dd7ccf53da3a9e"} Dec 03 00:49:26 crc kubenswrapper[4912]: I1203 00:49:26.931410 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" event={"ID":"26f3f384-6499-4824-8c10-de6382e27d8b","Type":"ContainerStarted","Data":"7cc385ce6ba75ac85fbbee14f194beb9d3255cd4125a2b8d83bd5473791a2a5d"} Dec 03 00:49:27 crc kubenswrapper[4912]: I1203 00:49:27.006647 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22763c05-583a-4681-bbf4-882fd2a8d7ed","Type":"ContainerStarted","Data":"0caf1d4d8b9be3f1786fb42971254e26ad6101685c058f7cb27639349f08d83c"} Dec 03 00:49:27 crc kubenswrapper[4912]: I1203 00:49:27.156967 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:27 crc kubenswrapper[4912]: W1203 00:49:27.311405 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaa68042_1653_4df5_9249_6263d3fcbd52.slice/crio-1bb5ea542a2cd85995a933b9671dde2cb205c71e99ab775e919d1979316bbcdb WatchSource:0}: Error finding container 1bb5ea542a2cd85995a933b9671dde2cb205c71e99ab775e919d1979316bbcdb: Status 404 returned error can't find the container with id 1bb5ea542a2cd85995a933b9671dde2cb205c71e99ab775e919d1979316bbcdb Dec 03 00:49:27 crc kubenswrapper[4912]: I1203 00:49:27.348162 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:27 crc kubenswrapper[4912]: I1203 00:49:27.476855 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75f549cc78-m5rcp" Dec 03 00:49:27 crc kubenswrapper[4912]: I1203 00:49:27.573009 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-95b564956-7sj8f"] Dec 03 00:49:27 crc kubenswrapper[4912]: I1203 00:49:27.573281 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-95b564956-7sj8f" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api-log" containerID="cri-o://f022956c3e095d4cbedf0a590c2565759de5d9f1252e632d331d2c7aa65ff179" gracePeriod=30 Dec 03 00:49:27 crc kubenswrapper[4912]: I1203 00:49:27.573916 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-95b564956-7sj8f" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api" containerID="cri-o://b199cb1e084a4ef25b9e2754a30d6236eeb6172edd6f45aae0e63b28d65866f7" gracePeriod=30 Dec 03 00:49:27 crc kubenswrapper[4912]: I1203 00:49:27.616031 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-95b564956-7sj8f" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.182:9311/healthcheck\": EOF" Dec 03 00:49:28 crc kubenswrapper[4912]: I1203 00:49:28.074022 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerStarted","Data":"1bb5ea542a2cd85995a933b9671dde2cb205c71e99ab775e919d1979316bbcdb"} Dec 03 00:49:28 crc kubenswrapper[4912]: I1203 00:49:28.104697 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 00:49:28 crc kubenswrapper[4912]: I1203 00:49:28.133493 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22763c05-583a-4681-bbf4-882fd2a8d7ed","Type":"ContainerStarted","Data":"0c4c3865b6e4d25992684d34e8ec5c5a1579024e07c7418c6c44b071c311be45"} Dec 03 00:49:28 crc kubenswrapper[4912]: I1203 00:49:28.138621 4912 generic.go:334] "Generic (PLEG): container finished" podID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerID="f022956c3e095d4cbedf0a590c2565759de5d9f1252e632d331d2c7aa65ff179" exitCode=143 Dec 03 00:49:28 crc kubenswrapper[4912]: I1203 00:49:28.138704 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-95b564956-7sj8f" event={"ID":"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e","Type":"ContainerDied","Data":"f022956c3e095d4cbedf0a590c2565759de5d9f1252e632d331d2c7aa65ff179"} Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.163834 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa","Type":"ContainerStarted","Data":"86a0c50b00cc8e42b594c06f1c02e1f336f438f40c70e399a8750d09a0ba06ca"} Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.168638 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" event={"ID":"26f3f384-6499-4824-8c10-de6382e27d8b","Type":"ContainerStarted","Data":"814304fa1d3f48577aeb5a367cdc99bdd58f1b535c6730312645d059b030a7ec"} Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.168808 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.177063 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerStarted","Data":"1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1"} Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.185617 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22763c05-583a-4681-bbf4-882fd2a8d7ed","Type":"ContainerStarted","Data":"eb2afd092785dbdfb95cd0b170a4e990e35a2115166ac5a87cf5a04eb82c43a5"} Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.185791 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerName="cinder-api-log" containerID="cri-o://0c4c3865b6e4d25992684d34e8ec5c5a1579024e07c7418c6c44b071c311be45" gracePeriod=30 Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.186046 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.186401 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerName="cinder-api" containerID="cri-o://eb2afd092785dbdfb95cd0b170a4e990e35a2115166ac5a87cf5a04eb82c43a5" gracePeriod=30 Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.221439 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" podStartSLOduration=4.221398675 podStartE2EDuration="4.221398675s" podCreationTimestamp="2025-12-03 00:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:29.185517716 +0000 UTC m=+1554.827538286" watchObservedRunningTime="2025-12-03 00:49:29.221398675 +0000 UTC m=+1554.863419235" Dec 03 00:49:29 crc kubenswrapper[4912]: I1203 00:49:29.230282 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.230259121 podStartE2EDuration="4.230259121s" podCreationTimestamp="2025-12-03 00:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:29.222207967 +0000 UTC m=+1554.864228527" watchObservedRunningTime="2025-12-03 00:49:29.230259121 +0000 UTC m=+1554.872279681" Dec 03 00:49:30 crc kubenswrapper[4912]: I1203 00:49:30.252746 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerStarted","Data":"1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428"} Dec 03 00:49:30 crc kubenswrapper[4912]: I1203 00:49:30.256499 4912 generic.go:334] "Generic (PLEG): container finished" podID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerID="0c4c3865b6e4d25992684d34e8ec5c5a1579024e07c7418c6c44b071c311be45" exitCode=143 Dec 03 00:49:30 crc kubenswrapper[4912]: I1203 00:49:30.258072 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22763c05-583a-4681-bbf4-882fd2a8d7ed","Type":"ContainerDied","Data":"0c4c3865b6e4d25992684d34e8ec5c5a1579024e07c7418c6c44b071c311be45"} Dec 03 00:49:31 crc kubenswrapper[4912]: I1203 00:49:31.275610 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerStarted","Data":"3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5"} Dec 03 00:49:31 crc kubenswrapper[4912]: I1203 00:49:31.282145 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa","Type":"ContainerStarted","Data":"0e745caaa6422c26ae778e229e667daf407ec5dd7d0a190c76a24af910a6d18f"} Dec 03 00:49:31 crc kubenswrapper[4912]: I1203 00:49:31.320153 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.04722213 podStartE2EDuration="6.320124164s" podCreationTimestamp="2025-12-03 00:49:25 +0000 UTC" firstStartedPulling="2025-12-03 00:49:26.122723501 +0000 UTC m=+1551.764744061" lastFinishedPulling="2025-12-03 00:49:27.395625535 +0000 UTC m=+1553.037646095" observedRunningTime="2025-12-03 00:49:31.311519955 +0000 UTC m=+1556.953540575" watchObservedRunningTime="2025-12-03 00:49:31.320124164 +0000 UTC m=+1556.962144724" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.085827 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-95b564956-7sj8f" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.182:9311/healthcheck\": read tcp 10.217.0.2:60650->10.217.0.182:9311: read: connection reset by peer" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.087364 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-95b564956-7sj8f" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.182:9311/healthcheck\": read tcp 10.217.0.2:60656->10.217.0.182:9311: read: connection reset by peer" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.310300 4912 generic.go:334] "Generic (PLEG): container finished" podID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerID="b199cb1e084a4ef25b9e2754a30d6236eeb6172edd6f45aae0e63b28d65866f7" exitCode=0 Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.310983 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-95b564956-7sj8f" event={"ID":"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e","Type":"ContainerDied","Data":"b199cb1e084a4ef25b9e2754a30d6236eeb6172edd6f45aae0e63b28d65866f7"} Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.596784 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.694796 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data-custom\") pod \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.695623 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6d6m\" (UniqueName: \"kubernetes.io/projected/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-kube-api-access-j6d6m\") pod \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.695707 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data\") pod \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.695821 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-logs\") pod \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.695929 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-combined-ca-bundle\") pod \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\" (UID: \"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e\") " Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.697235 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-logs" (OuterVolumeSpecName: "logs") pod "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" (UID: "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.703070 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-kube-api-access-j6d6m" (OuterVolumeSpecName: "kube-api-access-j6d6m") pod "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" (UID: "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e"). InnerVolumeSpecName "kube-api-access-j6d6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.728697 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" (UID: "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.745885 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" (UID: "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.798805 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data" (OuterVolumeSpecName: "config-data") pod "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" (UID: "11bf2e4c-b0d3-4ddf-89c6-6ab83386351e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.798856 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.798892 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6d6m\" (UniqueName: \"kubernetes.io/projected/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-kube-api-access-j6d6m\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.798905 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.798918 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.883553 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-59fd659dd5-vmcdc" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.901107 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.976538 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54f689d9c8-bxnjz"] Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.979150 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54f689d9c8-bxnjz" podUID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerName="neutron-api" containerID="cri-o://a8ae68d8961560dd4cb9207fc6ae82494366831638580b903327485478be6d64" gracePeriod=30 Dec 03 00:49:32 crc kubenswrapper[4912]: I1203 00:49:32.979238 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54f689d9c8-bxnjz" podUID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerName="neutron-httpd" containerID="cri-o://9f1699b7d36aca8216a0caf12a1d18f162ef66ab271b63e30b8ddea67b51299f" gracePeriod=30 Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.332009 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerStarted","Data":"b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc"} Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.332288 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.340747 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-95b564956-7sj8f" event={"ID":"11bf2e4c-b0d3-4ddf-89c6-6ab83386351e","Type":"ContainerDied","Data":"08b8ee6f7c345b4e9627df72004e8b0aa5145afd491a51670ee6f0a74a407821"} Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.340816 4912 scope.go:117] "RemoveContainer" containerID="b199cb1e084a4ef25b9e2754a30d6236eeb6172edd6f45aae0e63b28d65866f7" Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.340845 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-95b564956-7sj8f" Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.345471 4912 generic.go:334] "Generic (PLEG): container finished" podID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerID="9f1699b7d36aca8216a0caf12a1d18f162ef66ab271b63e30b8ddea67b51299f" exitCode=0 Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.345531 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f689d9c8-bxnjz" event={"ID":"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3","Type":"ContainerDied","Data":"9f1699b7d36aca8216a0caf12a1d18f162ef66ab271b63e30b8ddea67b51299f"} Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.400021 4912 scope.go:117] "RemoveContainer" containerID="f022956c3e095d4cbedf0a590c2565759de5d9f1252e632d331d2c7aa65ff179" Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.400019 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.319468867 podStartE2EDuration="8.39999139s" podCreationTimestamp="2025-12-03 00:49:25 +0000 UTC" firstStartedPulling="2025-12-03 00:49:27.316376068 +0000 UTC m=+1552.958396618" lastFinishedPulling="2025-12-03 00:49:32.396898581 +0000 UTC m=+1558.038919141" observedRunningTime="2025-12-03 00:49:33.364324278 +0000 UTC m=+1559.006344858" watchObservedRunningTime="2025-12-03 00:49:33.39999139 +0000 UTC m=+1559.042011950" Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.415752 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-95b564956-7sj8f"] Dec 03 00:49:33 crc kubenswrapper[4912]: I1203 00:49:33.427094 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-95b564956-7sj8f"] Dec 03 00:49:34 crc kubenswrapper[4912]: I1203 00:49:34.595798 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" path="/var/lib/kubelet/pods/11bf2e4c-b0d3-4ddf-89c6-6ab83386351e/volumes" Dec 03 00:49:35 crc kubenswrapper[4912]: I1203 00:49:35.518502 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 00:49:35 crc kubenswrapper[4912]: I1203 00:49:35.555937 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:49:35 crc kubenswrapper[4912]: I1203 00:49:35.659622 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tvc6r"] Dec 03 00:49:35 crc kubenswrapper[4912]: I1203 00:49:35.660235 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" podUID="2bc22778-372c-43ae-91cb-1007a3163d0d" containerName="dnsmasq-dns" containerID="cri-o://03ebca25b12667bf34b1085e1c7e3d2b91419437061ea1f4d17c9f80a2e3975c" gracePeriod=10 Dec 03 00:49:35 crc kubenswrapper[4912]: I1203 00:49:35.976582 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.417807 4912 generic.go:334] "Generic (PLEG): container finished" podID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerID="a8ae68d8961560dd4cb9207fc6ae82494366831638580b903327485478be6d64" exitCode=0 Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.418205 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f689d9c8-bxnjz" event={"ID":"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3","Type":"ContainerDied","Data":"a8ae68d8961560dd4cb9207fc6ae82494366831638580b903327485478be6d64"} Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.418242 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54f689d9c8-bxnjz" event={"ID":"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3","Type":"ContainerDied","Data":"434b65e23d30e8f8fbc3e6d469fc7d72f358238434a252dc6346edb0bd0c4ac4"} Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.418258 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="434b65e23d30e8f8fbc3e6d469fc7d72f358238434a252dc6346edb0bd0c4ac4" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.426194 4912 generic.go:334] "Generic (PLEG): container finished" podID="2bc22778-372c-43ae-91cb-1007a3163d0d" containerID="03ebca25b12667bf34b1085e1c7e3d2b91419437061ea1f4d17c9f80a2e3975c" exitCode=0 Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.426566 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" event={"ID":"2bc22778-372c-43ae-91cb-1007a3163d0d","Type":"ContainerDied","Data":"03ebca25b12667bf34b1085e1c7e3d2b91419437061ea1f4d17c9f80a2e3975c"} Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.446043 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.466168 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69d59dff98-dh972" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.478914 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.494934 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.563576 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.626360 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-swift-storage-0\") pod \"2bc22778-372c-43ae-91cb-1007a3163d0d\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.626766 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-config\") pod \"2bc22778-372c-43ae-91cb-1007a3163d0d\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.627012 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-ovndb-tls-certs\") pod \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.627187 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76whb\" (UniqueName: \"kubernetes.io/projected/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-kube-api-access-76whb\") pod \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.651585 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-httpd-config\") pod \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.652143 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-svc\") pod \"2bc22778-372c-43ae-91cb-1007a3163d0d\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.652242 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-nb\") pod \"2bc22778-372c-43ae-91cb-1007a3163d0d\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.652349 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-sb\") pod \"2bc22778-372c-43ae-91cb-1007a3163d0d\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.652529 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-config\") pod \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.652619 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrkxl\" (UniqueName: \"kubernetes.io/projected/2bc22778-372c-43ae-91cb-1007a3163d0d-kube-api-access-vrkxl\") pod \"2bc22778-372c-43ae-91cb-1007a3163d0d\" (UID: \"2bc22778-372c-43ae-91cb-1007a3163d0d\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.652707 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-combined-ca-bundle\") pod \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\" (UID: \"1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3\") " Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.661876 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" (UID: "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.695409 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-kube-api-access-76whb" (OuterVolumeSpecName: "kube-api-access-76whb") pod "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" (UID: "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3"). InnerVolumeSpecName "kube-api-access-76whb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.695362 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bc22778-372c-43ae-91cb-1007a3163d0d-kube-api-access-vrkxl" (OuterVolumeSpecName: "kube-api-access-vrkxl") pod "2bc22778-372c-43ae-91cb-1007a3163d0d" (UID: "2bc22778-372c-43ae-91cb-1007a3163d0d"). InnerVolumeSpecName "kube-api-access-vrkxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.733565 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2bc22778-372c-43ae-91cb-1007a3163d0d" (UID: "2bc22778-372c-43ae-91cb-1007a3163d0d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.758103 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrkxl\" (UniqueName: \"kubernetes.io/projected/2bc22778-372c-43ae-91cb-1007a3163d0d-kube-api-access-vrkxl\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.758811 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.758918 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76whb\" (UniqueName: \"kubernetes.io/projected/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-kube-api-access-76whb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.758991 4912 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.771186 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-config" (OuterVolumeSpecName: "config") pod "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" (UID: "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.818702 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2bc22778-372c-43ae-91cb-1007a3163d0d" (UID: "2bc22778-372c-43ae-91cb-1007a3163d0d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.827230 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2bc22778-372c-43ae-91cb-1007a3163d0d" (UID: "2bc22778-372c-43ae-91cb-1007a3163d0d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.856845 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" (UID: "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.861601 4912 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.861643 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.861654 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.861666 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.866414 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" (UID: "1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.877343 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2bc22778-372c-43ae-91cb-1007a3163d0d" (UID: "2bc22778-372c-43ae-91cb-1007a3163d0d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.886317 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-config" (OuterVolumeSpecName: "config") pod "2bc22778-372c-43ae-91cb-1007a3163d0d" (UID: "2bc22778-372c-43ae-91cb-1007a3163d0d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.963607 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.963653 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:36 crc kubenswrapper[4912]: I1203 00:49:36.963666 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2bc22778-372c-43ae-91cb-1007a3163d0d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.441338 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.441321 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tvc6r" event={"ID":"2bc22778-372c-43ae-91cb-1007a3163d0d","Type":"ContainerDied","Data":"c90742885165f06b83a62992922662062fd353699d1344ad1b5a8d30445cd432"} Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.441909 4912 scope.go:117] "RemoveContainer" containerID="03ebca25b12667bf34b1085e1c7e3d2b91419437061ea1f4d17c9f80a2e3975c" Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.441596 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerName="cinder-scheduler" containerID="cri-o://86a0c50b00cc8e42b594c06f1c02e1f336f438f40c70e399a8750d09a0ba06ca" gracePeriod=30 Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.442261 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerName="probe" containerID="cri-o://0e745caaa6422c26ae778e229e667daf407ec5dd7d0a190c76a24af910a6d18f" gracePeriod=30 Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.442265 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54f689d9c8-bxnjz" Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.487714 4912 scope.go:117] "RemoveContainer" containerID="cfb1e0086623df4dc2bbcf49985f4d74c93912cb4b30cdb6dde793d1f2a5c6c9" Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.529519 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tvc6r"] Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.549304 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tvc6r"] Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.564565 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54f689d9c8-bxnjz"] Dec 03 00:49:37 crc kubenswrapper[4912]: I1203 00:49:37.578809 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-54f689d9c8-bxnjz"] Dec 03 00:49:38 crc kubenswrapper[4912]: I1203 00:49:38.211755 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 00:49:38 crc kubenswrapper[4912]: I1203 00:49:38.473288 4912 generic.go:334] "Generic (PLEG): container finished" podID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerID="0e745caaa6422c26ae778e229e667daf407ec5dd7d0a190c76a24af910a6d18f" exitCode=0 Dec 03 00:49:38 crc kubenswrapper[4912]: I1203 00:49:38.473366 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa","Type":"ContainerDied","Data":"0e745caaa6422c26ae778e229e667daf407ec5dd7d0a190c76a24af910a6d18f"} Dec 03 00:49:38 crc kubenswrapper[4912]: I1203 00:49:38.592614 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" path="/var/lib/kubelet/pods/1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3/volumes" Dec 03 00:49:38 crc kubenswrapper[4912]: I1203 00:49:38.593733 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bc22778-372c-43ae-91cb-1007a3163d0d" path="/var/lib/kubelet/pods/2bc22778-372c-43ae-91cb-1007a3163d0d/volumes" Dec 03 00:49:40 crc kubenswrapper[4912]: I1203 00:49:40.504401 4912 generic.go:334] "Generic (PLEG): container finished" podID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerID="86a0c50b00cc8e42b594c06f1c02e1f336f438f40c70e399a8750d09a0ba06ca" exitCode=0 Dec 03 00:49:40 crc kubenswrapper[4912]: I1203 00:49:40.504484 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa","Type":"ContainerDied","Data":"86a0c50b00cc8e42b594c06f1c02e1f336f438f40c70e399a8750d09a0ba06ca"} Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.014631 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.069452 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-etc-machine-id\") pod \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.069523 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data-custom\") pod \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.069577 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-scripts\") pod \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.069612 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-combined-ca-bundle\") pod \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.069614 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" (UID: "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.069639 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfpmr\" (UniqueName: \"kubernetes.io/projected/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-kube-api-access-nfpmr\") pod \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.069812 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data\") pod \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\" (UID: \"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa\") " Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.070722 4912 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.077347 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-scripts" (OuterVolumeSpecName: "scripts") pod "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" (UID: "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.082463 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" (UID: "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.107883 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-kube-api-access-nfpmr" (OuterVolumeSpecName: "kube-api-access-nfpmr") pod "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" (UID: "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa"). InnerVolumeSpecName "kube-api-access-nfpmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.172894 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.172945 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.172960 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfpmr\" (UniqueName: \"kubernetes.io/projected/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-kube-api-access-nfpmr\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.212657 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" (UID: "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.241909 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data" (OuterVolumeSpecName: "config-data") pod "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" (UID: "affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.275421 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.275976 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.520932 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa","Type":"ContainerDied","Data":"c94935886c4e8dce7700d534d888ba822644d7b859b6b66d8deb6c8cc3470d0f"} Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.520992 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.521007 4912 scope.go:117] "RemoveContainer" containerID="0e745caaa6422c26ae778e229e667daf407ec5dd7d0a190c76a24af910a6d18f" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.547974 4912 scope.go:117] "RemoveContainer" containerID="86a0c50b00cc8e42b594c06f1c02e1f336f438f40c70e399a8750d09a0ba06ca" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.562258 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.570531 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.602938 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 00:49:41 crc kubenswrapper[4912]: E1203 00:49:41.603447 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc22778-372c-43ae-91cb-1007a3163d0d" containerName="dnsmasq-dns" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603479 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc22778-372c-43ae-91cb-1007a3163d0d" containerName="dnsmasq-dns" Dec 03 00:49:41 crc kubenswrapper[4912]: E1203 00:49:41.603523 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603532 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api" Dec 03 00:49:41 crc kubenswrapper[4912]: E1203 00:49:41.603542 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerName="neutron-httpd" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603550 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerName="neutron-httpd" Dec 03 00:49:41 crc kubenswrapper[4912]: E1203 00:49:41.603564 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerName="neutron-api" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603570 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerName="neutron-api" Dec 03 00:49:41 crc kubenswrapper[4912]: E1203 00:49:41.603583 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerName="probe" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603589 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerName="probe" Dec 03 00:49:41 crc kubenswrapper[4912]: E1203 00:49:41.603600 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerName="cinder-scheduler" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603606 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerName="cinder-scheduler" Dec 03 00:49:41 crc kubenswrapper[4912]: E1203 00:49:41.603615 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bc22778-372c-43ae-91cb-1007a3163d0d" containerName="init" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603622 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bc22778-372c-43ae-91cb-1007a3163d0d" containerName="init" Dec 03 00:49:41 crc kubenswrapper[4912]: E1203 00:49:41.603639 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api-log" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603647 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api-log" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603837 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603852 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerName="neutron-api" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603866 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="11bf2e4c-b0d3-4ddf-89c6-6ab83386351e" containerName="barbican-api-log" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603878 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerName="cinder-scheduler" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603891 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bc22778-372c-43ae-91cb-1007a3163d0d" containerName="dnsmasq-dns" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603904 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" containerName="probe" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.603918 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="1023ccf5-0ef7-4c98-ac78-3a9cf18dd2d3" containerName="neutron-httpd" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.605075 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.609202 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.617100 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.686728 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.686770 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.686813 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smb2v\" (UniqueName: \"kubernetes.io/projected/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-kube-api-access-smb2v\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.686923 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-scripts\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.686968 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.687002 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-config-data\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.738946 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-d9977975b-bq8m5" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.789196 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.789291 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-config-data\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.789462 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.789482 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.789509 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smb2v\" (UniqueName: \"kubernetes.io/projected/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-kube-api-access-smb2v\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.789550 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-scripts\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.792263 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.798184 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.798569 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-scripts\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.809906 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.810292 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-config-data\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.821373 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smb2v\" (UniqueName: \"kubernetes.io/projected/8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0-kube-api-access-smb2v\") pod \"cinder-scheduler-0\" (UID: \"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0\") " pod="openstack/cinder-scheduler-0" Dec 03 00:49:41 crc kubenswrapper[4912]: I1203 00:49:41.927578 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 00:49:42 crc kubenswrapper[4912]: I1203 00:49:42.550767 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 00:49:42 crc kubenswrapper[4912]: W1203 00:49:42.573723 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8af6e3fe_8a7a_480e_80b6_cd5d5e634fd0.slice/crio-8c6c5c17dbd7748c7a9da7dece93451b864d3cc580aeff6945d1e96490514bf5 WatchSource:0}: Error finding container 8c6c5c17dbd7748c7a9da7dece93451b864d3cc580aeff6945d1e96490514bf5: Status 404 returned error can't find the container with id 8c6c5c17dbd7748c7a9da7dece93451b864d3cc580aeff6945d1e96490514bf5 Dec 03 00:49:42 crc kubenswrapper[4912]: I1203 00:49:42.622070 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa" path="/var/lib/kubelet/pods/affbee0c-a6f2-4e2c-a7fe-c9fc2bd446fa/volumes" Dec 03 00:49:43 crc kubenswrapper[4912]: I1203 00:49:43.553199 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0","Type":"ContainerStarted","Data":"00662f873d66dc26ec6888fd9a93d7f819b9b43a57fca5ea9ff057ee74b84599"} Dec 03 00:49:43 crc kubenswrapper[4912]: I1203 00:49:43.553973 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0","Type":"ContainerStarted","Data":"8c6c5c17dbd7748c7a9da7dece93451b864d3cc580aeff6945d1e96490514bf5"} Dec 03 00:49:44 crc kubenswrapper[4912]: I1203 00:49:44.566251 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0","Type":"ContainerStarted","Data":"3e26f9b5f6c725cb087a974826523857c124a1bdf62f913781e48efe4689d1d7"} Dec 03 00:49:44 crc kubenswrapper[4912]: I1203 00:49:44.600740 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.600706111 podStartE2EDuration="3.600706111s" podCreationTimestamp="2025-12-03 00:49:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:44.592949375 +0000 UTC m=+1570.234969945" watchObservedRunningTime="2025-12-03 00:49:44.600706111 +0000 UTC m=+1570.242726671" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.386882 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7ddd66b7bc-vwpcg"] Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.389675 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.391610 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.392276 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.394034 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.427142 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7ddd66b7bc-vwpcg"] Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.514001 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/540979f8-98e6-40f8-9f82-1e7b924d984e-etc-swift\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.514076 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-combined-ca-bundle\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.514145 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/540979f8-98e6-40f8-9f82-1e7b924d984e-run-httpd\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.514180 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-public-tls-certs\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.514203 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-internal-tls-certs\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.514246 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvjjk\" (UniqueName: \"kubernetes.io/projected/540979f8-98e6-40f8-9f82-1e7b924d984e-kube-api-access-jvjjk\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.514267 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-config-data\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.514312 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/540979f8-98e6-40f8-9f82-1e7b924d984e-log-httpd\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.618074 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/540979f8-98e6-40f8-9f82-1e7b924d984e-log-httpd\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.618676 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/540979f8-98e6-40f8-9f82-1e7b924d984e-log-httpd\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.618718 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/540979f8-98e6-40f8-9f82-1e7b924d984e-etc-swift\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.618886 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-combined-ca-bundle\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.619186 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/540979f8-98e6-40f8-9f82-1e7b924d984e-run-httpd\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.620156 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/540979f8-98e6-40f8-9f82-1e7b924d984e-run-httpd\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.620264 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-public-tls-certs\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.620364 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-internal-tls-certs\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.620523 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvjjk\" (UniqueName: \"kubernetes.io/projected/540979f8-98e6-40f8-9f82-1e7b924d984e-kube-api-access-jvjjk\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.620650 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-config-data\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.630252 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-internal-tls-certs\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.631250 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-config-data\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.632342 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-public-tls-certs\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.632675 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/540979f8-98e6-40f8-9f82-1e7b924d984e-combined-ca-bundle\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.634679 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/540979f8-98e6-40f8-9f82-1e7b924d984e-etc-swift\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.648719 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvjjk\" (UniqueName: \"kubernetes.io/projected/540979f8-98e6-40f8-9f82-1e7b924d984e-kube-api-access-jvjjk\") pod \"swift-proxy-7ddd66b7bc-vwpcg\" (UID: \"540979f8-98e6-40f8-9f82-1e7b924d984e\") " pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.717879 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.900139 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.902774 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.908588 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.908828 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-q27v4" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.909001 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.929350 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.932543 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsxr7\" (UniqueName: \"kubernetes.io/projected/be4584b5-a436-42e3-8ced-da24fb252c2f-kube-api-access-jsxr7\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.932601 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config-secret\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.932719 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.932855 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:46 crc kubenswrapper[4912]: I1203 00:49:46.938520 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.035817 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.035913 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsxr7\" (UniqueName: \"kubernetes.io/projected/be4584b5-a436-42e3-8ced-da24fb252c2f-kube-api-access-jsxr7\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.035970 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config-secret\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.036130 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.037023 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.057261 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config-secret\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.063764 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-combined-ca-bundle\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.064889 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsxr7\" (UniqueName: \"kubernetes.io/projected/be4584b5-a436-42e3-8ced-da24fb252c2f-kube-api-access-jsxr7\") pod \"openstackclient\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.141579 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.142289 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.194394 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.208349 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.211159 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.230191 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.253024 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1cf8990-bf85-4de8-8121-73c43a7eccd9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.253100 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b1cf8990-bf85-4de8-8121-73c43a7eccd9-openstack-config\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.253235 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b1cf8990-bf85-4de8-8121-73c43a7eccd9-openstack-config-secret\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.253366 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttpbs\" (UniqueName: \"kubernetes.io/projected/b1cf8990-bf85-4de8-8121-73c43a7eccd9-kube-api-access-ttpbs\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: E1203 00:49:47.294753 4912 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 03 00:49:47 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_be4584b5-a436-42e3-8ced-da24fb252c2f_0(534119fe88c27817653f3302b22519efe50a69d678312ad366abfc98f106dcd6): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"534119fe88c27817653f3302b22519efe50a69d678312ad366abfc98f106dcd6" Netns:"/var/run/netns/6cf0dfd1-b39f-4408-a581-2bc483663ee2" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=534119fe88c27817653f3302b22519efe50a69d678312ad366abfc98f106dcd6;K8S_POD_UID=be4584b5-a436-42e3-8ced-da24fb252c2f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/be4584b5-a436-42e3-8ced-da24fb252c2f]: expected pod UID "be4584b5-a436-42e3-8ced-da24fb252c2f" but got "b1cf8990-bf85-4de8-8121-73c43a7eccd9" from Kube API Dec 03 00:49:47 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:49:47 crc kubenswrapper[4912]: > Dec 03 00:49:47 crc kubenswrapper[4912]: E1203 00:49:47.294869 4912 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 03 00:49:47 crc kubenswrapper[4912]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_be4584b5-a436-42e3-8ced-da24fb252c2f_0(534119fe88c27817653f3302b22519efe50a69d678312ad366abfc98f106dcd6): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"534119fe88c27817653f3302b22519efe50a69d678312ad366abfc98f106dcd6" Netns:"/var/run/netns/6cf0dfd1-b39f-4408-a581-2bc483663ee2" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=534119fe88c27817653f3302b22519efe50a69d678312ad366abfc98f106dcd6;K8S_POD_UID=be4584b5-a436-42e3-8ced-da24fb252c2f" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/be4584b5-a436-42e3-8ced-da24fb252c2f]: expected pod UID "be4584b5-a436-42e3-8ced-da24fb252c2f" but got "b1cf8990-bf85-4de8-8121-73c43a7eccd9" from Kube API Dec 03 00:49:47 crc kubenswrapper[4912]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 03 00:49:47 crc kubenswrapper[4912]: > pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: W1203 00:49:47.341934 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod540979f8_98e6_40f8_9f82_1e7b924d984e.slice/crio-99bb63bb6b1e45021140a31f00dd81358e2226bfe5eb17864b9a7b2bbeee31f6 WatchSource:0}: Error finding container 99bb63bb6b1e45021140a31f00dd81358e2226bfe5eb17864b9a7b2bbeee31f6: Status 404 returned error can't find the container with id 99bb63bb6b1e45021140a31f00dd81358e2226bfe5eb17864b9a7b2bbeee31f6 Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.342294 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7ddd66b7bc-vwpcg"] Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.355964 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1cf8990-bf85-4de8-8121-73c43a7eccd9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.356034 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b1cf8990-bf85-4de8-8121-73c43a7eccd9-openstack-config\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.356262 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b1cf8990-bf85-4de8-8121-73c43a7eccd9-openstack-config-secret\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.356989 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttpbs\" (UniqueName: \"kubernetes.io/projected/b1cf8990-bf85-4de8-8121-73c43a7eccd9-kube-api-access-ttpbs\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.357264 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b1cf8990-bf85-4de8-8121-73c43a7eccd9-openstack-config\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.362523 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b1cf8990-bf85-4de8-8121-73c43a7eccd9-openstack-config-secret\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.363140 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1cf8990-bf85-4de8-8121-73c43a7eccd9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.377962 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttpbs\" (UniqueName: \"kubernetes.io/projected/b1cf8990-bf85-4de8-8121-73c43a7eccd9-kube-api-access-ttpbs\") pod \"openstackclient\" (UID: \"b1cf8990-bf85-4de8-8121-73c43a7eccd9\") " pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.570355 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.641408 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.642662 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" event={"ID":"540979f8-98e6-40f8-9f82-1e7b924d984e","Type":"ContainerStarted","Data":"99bb63bb6b1e45021140a31f00dd81358e2226bfe5eb17864b9a7b2bbeee31f6"} Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.646314 4912 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="be4584b5-a436-42e3-8ced-da24fb252c2f" podUID="b1cf8990-bf85-4de8-8121-73c43a7eccd9" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.725459 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.868103 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config-secret\") pod \"be4584b5-a436-42e3-8ced-da24fb252c2f\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.868155 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config\") pod \"be4584b5-a436-42e3-8ced-da24fb252c2f\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.868309 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsxr7\" (UniqueName: \"kubernetes.io/projected/be4584b5-a436-42e3-8ced-da24fb252c2f-kube-api-access-jsxr7\") pod \"be4584b5-a436-42e3-8ced-da24fb252c2f\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.868464 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-combined-ca-bundle\") pod \"be4584b5-a436-42e3-8ced-da24fb252c2f\" (UID: \"be4584b5-a436-42e3-8ced-da24fb252c2f\") " Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.869110 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "be4584b5-a436-42e3-8ced-da24fb252c2f" (UID: "be4584b5-a436-42e3-8ced-da24fb252c2f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.875368 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "be4584b5-a436-42e3-8ced-da24fb252c2f" (UID: "be4584b5-a436-42e3-8ced-da24fb252c2f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.876391 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be4584b5-a436-42e3-8ced-da24fb252c2f-kube-api-access-jsxr7" (OuterVolumeSpecName: "kube-api-access-jsxr7") pod "be4584b5-a436-42e3-8ced-da24fb252c2f" (UID: "be4584b5-a436-42e3-8ced-da24fb252c2f"). InnerVolumeSpecName "kube-api-access-jsxr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.878403 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be4584b5-a436-42e3-8ced-da24fb252c2f" (UID: "be4584b5-a436-42e3-8ced-da24fb252c2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.970831 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.971326 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.971344 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be4584b5-a436-42e3-8ced-da24fb252c2f-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:47 crc kubenswrapper[4912]: I1203 00:49:47.971359 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsxr7\" (UniqueName: \"kubernetes.io/projected/be4584b5-a436-42e3-8ced-da24fb252c2f-kube-api-access-jsxr7\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.077548 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.077643 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.077710 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.078779 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.078854 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" gracePeriod=600 Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.149670 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 00:49:48 crc kubenswrapper[4912]: W1203 00:49:48.153645 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1cf8990_bf85_4de8_8121_73c43a7eccd9.slice/crio-76ab0ecda47763793d21aa4cdc8db3d7a09af73312d4ed3595a91cd965deb8c8 WatchSource:0}: Error finding container 76ab0ecda47763793d21aa4cdc8db3d7a09af73312d4ed3595a91cd965deb8c8: Status 404 returned error can't find the container with id 76ab0ecda47763793d21aa4cdc8db3d7a09af73312d4ed3595a91cd965deb8c8 Dec 03 00:49:48 crc kubenswrapper[4912]: E1203 00:49:48.210265 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.238031 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-vk882"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.239783 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.257240 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vk882"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.352101 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-4b7wk"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.354788 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.377781 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-9697-account-create-update-chh6n"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.379675 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.381478 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn4xm\" (UniqueName: \"kubernetes.io/projected/4cea1b95-ee14-4157-80f0-8785715d76fb-kube-api-access-gn4xm\") pod \"nova-api-db-create-vk882\" (UID: \"4cea1b95-ee14-4157-80f0-8785715d76fb\") " pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.381588 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cea1b95-ee14-4157-80f0-8785715d76fb-operator-scripts\") pod \"nova-api-db-create-vk882\" (UID: \"4cea1b95-ee14-4157-80f0-8785715d76fb\") " pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.384033 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.420265 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4b7wk"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.435497 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9697-account-create-update-chh6n"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.483928 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e0fc17b-4a05-4eef-87d9-1269183df954-operator-scripts\") pod \"nova-cell0-db-create-4b7wk\" (UID: \"5e0fc17b-4a05-4eef-87d9-1269183df954\") " pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.484012 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn4xm\" (UniqueName: \"kubernetes.io/projected/4cea1b95-ee14-4157-80f0-8785715d76fb-kube-api-access-gn4xm\") pod \"nova-api-db-create-vk882\" (UID: \"4cea1b95-ee14-4157-80f0-8785715d76fb\") " pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.484046 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-operator-scripts\") pod \"nova-api-9697-account-create-update-chh6n\" (UID: \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\") " pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.484084 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs6lb\" (UniqueName: \"kubernetes.io/projected/5e0fc17b-4a05-4eef-87d9-1269183df954-kube-api-access-bs6lb\") pod \"nova-cell0-db-create-4b7wk\" (UID: \"5e0fc17b-4a05-4eef-87d9-1269183df954\") " pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.484161 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cea1b95-ee14-4157-80f0-8785715d76fb-operator-scripts\") pod \"nova-api-db-create-vk882\" (UID: \"4cea1b95-ee14-4157-80f0-8785715d76fb\") " pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.484213 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgrm6\" (UniqueName: \"kubernetes.io/projected/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-kube-api-access-xgrm6\") pod \"nova-api-9697-account-create-update-chh6n\" (UID: \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\") " pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.485262 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cea1b95-ee14-4157-80f0-8785715d76fb-operator-scripts\") pod \"nova-api-db-create-vk882\" (UID: \"4cea1b95-ee14-4157-80f0-8785715d76fb\") " pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.515154 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn4xm\" (UniqueName: \"kubernetes.io/projected/4cea1b95-ee14-4157-80f0-8785715d76fb-kube-api-access-gn4xm\") pod \"nova-api-db-create-vk882\" (UID: \"4cea1b95-ee14-4157-80f0-8785715d76fb\") " pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.536694 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-24b5-account-create-update-4glzv"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.539915 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.543402 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.591270 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.604408 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be4584b5-a436-42e3-8ced-da24fb252c2f" path="/var/lib/kubelet/pods/be4584b5-a436-42e3-8ced-da24fb252c2f/volumes" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.605377 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-fqvcx"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.605761 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e0fc17b-4a05-4eef-87d9-1269183df954-operator-scripts\") pod \"nova-cell0-db-create-4b7wk\" (UID: \"5e0fc17b-4a05-4eef-87d9-1269183df954\") " pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.605936 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-operator-scripts\") pod \"nova-api-9697-account-create-update-chh6n\" (UID: \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\") " pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.606024 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs6lb\" (UniqueName: \"kubernetes.io/projected/5e0fc17b-4a05-4eef-87d9-1269183df954-kube-api-access-bs6lb\") pod \"nova-cell0-db-create-4b7wk\" (UID: \"5e0fc17b-4a05-4eef-87d9-1269183df954\") " pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.606295 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgrm6\" (UniqueName: \"kubernetes.io/projected/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-kube-api-access-xgrm6\") pod \"nova-api-9697-account-create-update-chh6n\" (UID: \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\") " pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.622358 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e0fc17b-4a05-4eef-87d9-1269183df954-operator-scripts\") pod \"nova-cell0-db-create-4b7wk\" (UID: \"5e0fc17b-4a05-4eef-87d9-1269183df954\") " pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.623172 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-operator-scripts\") pod \"nova-api-9697-account-create-update-chh6n\" (UID: \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\") " pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.623464 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.627316 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs6lb\" (UniqueName: \"kubernetes.io/projected/5e0fc17b-4a05-4eef-87d9-1269183df954-kube-api-access-bs6lb\") pod \"nova-cell0-db-create-4b7wk\" (UID: \"5e0fc17b-4a05-4eef-87d9-1269183df954\") " pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.640014 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-24b5-account-create-update-4glzv"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.646109 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgrm6\" (UniqueName: \"kubernetes.io/projected/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-kube-api-access-xgrm6\") pod \"nova-api-9697-account-create-update-chh6n\" (UID: \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\") " pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.659488 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fqvcx"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.691620 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.704070 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b1cf8990-bf85-4de8-8121-73c43a7eccd9","Type":"ContainerStarted","Data":"76ab0ecda47763793d21aa4cdc8db3d7a09af73312d4ed3595a91cd965deb8c8"} Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.709460 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.711478 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d27d\" (UniqueName: \"kubernetes.io/projected/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-kube-api-access-4d27d\") pod \"nova-cell1-db-create-fqvcx\" (UID: \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\") " pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.711559 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-operator-scripts\") pod \"nova-cell0-24b5-account-create-update-4glzv\" (UID: \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\") " pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.711682 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncbtk\" (UniqueName: \"kubernetes.io/projected/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-kube-api-access-ncbtk\") pod \"nova-cell0-24b5-account-create-update-4glzv\" (UID: \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\") " pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.711728 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-operator-scripts\") pod \"nova-cell1-db-create-fqvcx\" (UID: \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\") " pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.725809 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" event={"ID":"540979f8-98e6-40f8-9f82-1e7b924d984e","Type":"ContainerStarted","Data":"327c0c8bdc57197998c6d0188d5b2436d4c1aa9084509dd480a1a5de2ab86f27"} Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.725871 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" event={"ID":"540979f8-98e6-40f8-9f82-1e7b924d984e","Type":"ContainerStarted","Data":"1012dde8a329c3ad1512ca533241320256b582538f433f2267b73939732aa971"} Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.726519 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.726569 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.783139 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" exitCode=0 Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.783586 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.785410 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a"} Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.785470 4912 scope.go:117] "RemoveContainer" containerID="5d52c01534a739fd57a8ce1b61cd4e126b48cef8755e1aae547590b36373a992" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.786680 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:49:48 crc kubenswrapper[4912]: E1203 00:49:48.786951 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.791452 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-81fe-account-create-update-8dcm2"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.796044 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.803239 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.808216 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-81fe-account-create-update-8dcm2"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.814967 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d27d\" (UniqueName: \"kubernetes.io/projected/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-kube-api-access-4d27d\") pod \"nova-cell1-db-create-fqvcx\" (UID: \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\") " pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.815098 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-operator-scripts\") pod \"nova-cell0-24b5-account-create-update-4glzv\" (UID: \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\") " pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.815211 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncbtk\" (UniqueName: \"kubernetes.io/projected/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-kube-api-access-ncbtk\") pod \"nova-cell0-24b5-account-create-update-4glzv\" (UID: \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\") " pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.815254 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-operator-scripts\") pod \"nova-cell1-db-create-fqvcx\" (UID: \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\") " pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.814915 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" podStartSLOduration=2.814894017 podStartE2EDuration="2.814894017s" podCreationTimestamp="2025-12-03 00:49:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:48.783069047 +0000 UTC m=+1574.425089627" watchObservedRunningTime="2025-12-03 00:49:48.814894017 +0000 UTC m=+1574.456914577" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.817803 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-operator-scripts\") pod \"nova-cell0-24b5-account-create-update-4glzv\" (UID: \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\") " pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.828632 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-operator-scripts\") pod \"nova-cell1-db-create-fqvcx\" (UID: \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\") " pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.887228 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncbtk\" (UniqueName: \"kubernetes.io/projected/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-kube-api-access-ncbtk\") pod \"nova-cell0-24b5-account-create-update-4glzv\" (UID: \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\") " pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.892373 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d27d\" (UniqueName: \"kubernetes.io/projected/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-kube-api-access-4d27d\") pod \"nova-cell1-db-create-fqvcx\" (UID: \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\") " pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.919502 4912 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="be4584b5-a436-42e3-8ced-da24fb252c2f" podUID="b1cf8990-bf85-4de8-8121-73c43a7eccd9" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.921300 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-operator-scripts\") pod \"nova-cell1-81fe-account-create-update-8dcm2\" (UID: \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\") " pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.921400 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxs8t\" (UniqueName: \"kubernetes.io/projected/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-kube-api-access-vxs8t\") pod \"nova-cell1-81fe-account-create-update-8dcm2\" (UID: \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\") " pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.929876 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.951767 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.952215 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="ceilometer-central-agent" containerID="cri-o://1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1" gracePeriod=30 Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.952352 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="proxy-httpd" containerID="cri-o://b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc" gracePeriod=30 Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.952387 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="sg-core" containerID="cri-o://3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5" gracePeriod=30 Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.952443 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="ceilometer-notification-agent" containerID="cri-o://1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428" gracePeriod=30 Dec 03 00:49:48 crc kubenswrapper[4912]: I1203 00:49:48.964276 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.024002 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-operator-scripts\") pod \"nova-cell1-81fe-account-create-update-8dcm2\" (UID: \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\") " pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.024100 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxs8t\" (UniqueName: \"kubernetes.io/projected/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-kube-api-access-vxs8t\") pod \"nova-cell1-81fe-account-create-update-8dcm2\" (UID: \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\") " pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.025479 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-operator-scripts\") pod \"nova-cell1-81fe-account-create-update-8dcm2\" (UID: \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\") " pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.072969 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxs8t\" (UniqueName: \"kubernetes.io/projected/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-kube-api-access-vxs8t\") pod \"nova-cell1-81fe-account-create-update-8dcm2\" (UID: \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\") " pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.164005 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.244065 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.446754 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-vk882"] Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.690283 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4b7wk"] Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.715797 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-9697-account-create-update-chh6n"] Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.828698 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vk882" event={"ID":"4cea1b95-ee14-4157-80f0-8785715d76fb","Type":"ContainerStarted","Data":"a015f44208388f30316e6cffc96be4e6fa46ff7ea1559617ad575f6e7d07244e"} Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.841978 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9697-account-create-update-chh6n" event={"ID":"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72","Type":"ContainerStarted","Data":"bfafce7c2451377b8deb76263144d935b633fda987c7bce313f0e59fabbf57d9"} Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.849671 4912 generic.go:334] "Generic (PLEG): container finished" podID="caa68042-1653-4df5-9249-6263d3fcbd52" containerID="b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc" exitCode=0 Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.849705 4912 generic.go:334] "Generic (PLEG): container finished" podID="caa68042-1653-4df5-9249-6263d3fcbd52" containerID="3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5" exitCode=2 Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.849724 4912 generic.go:334] "Generic (PLEG): container finished" podID="caa68042-1653-4df5-9249-6263d3fcbd52" containerID="1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1" exitCode=0 Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.849758 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerDied","Data":"b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc"} Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.849783 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerDied","Data":"3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5"} Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.849792 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerDied","Data":"1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1"} Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.856919 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4b7wk" event={"ID":"5e0fc17b-4a05-4eef-87d9-1269183df954","Type":"ContainerStarted","Data":"338b4f9ad38150948bd08761d1218ac1fbc5246e68644371441037bcb9b402e9"} Dec 03 00:49:49 crc kubenswrapper[4912]: I1203 00:49:49.905732 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fqvcx"] Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.126613 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-81fe-account-create-update-8dcm2"] Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.150204 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-24b5-account-create-update-4glzv"] Dec 03 00:49:50 crc kubenswrapper[4912]: E1203 00:49:50.771593 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f8ba108_0d7f_4399_a95f_aaf6b65b3499.slice/crio-6bc31f24cf6fb25378a406779e58e5e45a75472810588f9b4e613dadd172fdd5.scope\": RecentStats: unable to find data in memory cache]" Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.961955 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-78c76df587-7nwws"] Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.969036 4912 generic.go:334] "Generic (PLEG): container finished" podID="3d9ac0c8-42e7-4d85-be36-fd702a7c8787" containerID="be579efe3bbae50445d1e5c57ef8a64c44ba098f9355fcb4f7d3fee15461a1a9" exitCode=0 Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.972053 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" event={"ID":"3d9ac0c8-42e7-4d85-be36-fd702a7c8787","Type":"ContainerDied","Data":"be579efe3bbae50445d1e5c57ef8a64c44ba098f9355fcb4f7d3fee15461a1a9"} Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.972142 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" event={"ID":"3d9ac0c8-42e7-4d85-be36-fd702a7c8787","Type":"ContainerStarted","Data":"cc4501acf377a3d00632a15bfb1de5cf8d369be6dcb4f3bd6f9c19d223cdf7fa"} Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.972250 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.979287 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-7lc7t" Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.981832 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.982273 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 03 00:49:50 crc kubenswrapper[4912]: I1203 00:49:50.991181 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-78c76df587-7nwws"] Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.017920 4912 generic.go:334] "Generic (PLEG): container finished" podID="4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9" containerID="41d342ee08eaa27f5b4fa2489afbec5b1bc97e532fc15e9f47b0efe8a1e3645f" exitCode=0 Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.018269 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-24b5-account-create-update-4glzv" event={"ID":"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9","Type":"ContainerDied","Data":"41d342ee08eaa27f5b4fa2489afbec5b1bc97e532fc15e9f47b0efe8a1e3645f"} Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.018373 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-24b5-account-create-update-4glzv" event={"ID":"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9","Type":"ContainerStarted","Data":"fee9728e66ce8ecbcdd92c7131260f22a94403c713f3f202285e76a8337318cb"} Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.030528 4912 generic.go:334] "Generic (PLEG): container finished" podID="8f8ba108-0d7f-4399-a95f-aaf6b65b3499" containerID="6bc31f24cf6fb25378a406779e58e5e45a75472810588f9b4e613dadd172fdd5" exitCode=0 Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.030645 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fqvcx" event={"ID":"8f8ba108-0d7f-4399-a95f-aaf6b65b3499","Type":"ContainerDied","Data":"6bc31f24cf6fb25378a406779e58e5e45a75472810588f9b4e613dadd172fdd5"} Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.030680 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fqvcx" event={"ID":"8f8ba108-0d7f-4399-a95f-aaf6b65b3499","Type":"ContainerStarted","Data":"6ee6a2720f17577561427c70e6e6e7cef36df3e6a2eea46e12953f11cef0ff9f"} Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.038957 4912 generic.go:334] "Generic (PLEG): container finished" podID="5e0fc17b-4a05-4eef-87d9-1269183df954" containerID="f2326a066994c7938c4e11983325cefb03f174cff70bdfaf1cf803dbeca640eb" exitCode=0 Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.039051 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4b7wk" event={"ID":"5e0fc17b-4a05-4eef-87d9-1269183df954","Type":"ContainerDied","Data":"f2326a066994c7938c4e11983325cefb03f174cff70bdfaf1cf803dbeca640eb"} Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.058474 4912 generic.go:334] "Generic (PLEG): container finished" podID="4cea1b95-ee14-4157-80f0-8785715d76fb" containerID="9396fe1c1097427efa3529e92e45ad6c7b7495250877f84663531e11bb6929d5" exitCode=0 Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.058862 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vk882" event={"ID":"4cea1b95-ee14-4157-80f0-8785715d76fb","Type":"ContainerDied","Data":"9396fe1c1097427efa3529e92e45ad6c7b7495250877f84663531e11bb6929d5"} Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.077603 4912 generic.go:334] "Generic (PLEG): container finished" podID="ae1a6a50-b1c5-43b9-994c-7d3072dd8f72" containerID="9543c465d05ab8f926067fdf27ccce5b142898b3c761307e8731c643f36dcb34" exitCode=0 Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.077669 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9697-account-create-update-chh6n" event={"ID":"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72","Type":"ContainerDied","Data":"9543c465d05ab8f926067fdf27ccce5b142898b3c761307e8731c643f36dcb34"} Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.091763 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.091860 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-combined-ca-bundle\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.091953 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data-custom\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.092037 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr28t\" (UniqueName: \"kubernetes.io/projected/4881ee51-d772-4954-8e62-29584cc7b3c8-kube-api-access-zr28t\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.167215 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-g7nr8"] Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.170693 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.194299 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data-custom\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.204089 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr28t\" (UniqueName: \"kubernetes.io/projected/4881ee51-d772-4954-8e62-29584cc7b3c8-kube-api-access-zr28t\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.204401 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.204567 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-combined-ca-bundle\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.210885 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data-custom\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.230618 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.231229 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-combined-ca-bundle\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.259778 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-g7nr8"] Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.299289 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr28t\" (UniqueName: \"kubernetes.io/projected/4881ee51-d772-4954-8e62-29584cc7b3c8-kube-api-access-zr28t\") pod \"heat-engine-78c76df587-7nwws\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.309387 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.309494 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44zd9\" (UniqueName: \"kubernetes.io/projected/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-kube-api-access-44zd9\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.309570 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.309625 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.309676 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-config\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.309737 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.330122 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.399119 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6d4656b7b7-t62gs"] Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.401132 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.420271 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.426200 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.426274 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44zd9\" (UniqueName: \"kubernetes.io/projected/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-kube-api-access-44zd9\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.426393 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.426509 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.426606 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-config\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.426708 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.427824 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.428340 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.429239 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.437669 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-config\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.438552 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.466808 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44zd9\" (UniqueName: \"kubernetes.io/projected/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-kube-api-access-44zd9\") pod \"dnsmasq-dns-688b9f5b49-g7nr8\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.547503 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-combined-ca-bundle\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.547643 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.547773 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data-custom\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.547863 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9wm9\" (UniqueName: \"kubernetes.io/projected/ef6b7b00-8f47-4d05-92f0-b43367f92c12-kube-api-access-l9wm9\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.548943 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6d4656b7b7-t62gs"] Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.633474 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-55bc76fdd7-jqgvd"] Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.635161 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.643720 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.651082 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data-custom\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.651204 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9wm9\" (UniqueName: \"kubernetes.io/projected/ef6b7b00-8f47-4d05-92f0-b43367f92c12-kube-api-access-l9wm9\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.651250 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-combined-ca-bundle\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.651305 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.704545 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data-custom\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.710231 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-combined-ca-bundle\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.711703 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.718064 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.761504 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.761626 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-combined-ca-bundle\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.761684 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnnhb\" (UniqueName: \"kubernetes.io/projected/5712fffb-e277-4ebd-a130-fee6f21273dc-kube-api-access-hnnhb\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.761759 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data-custom\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.773813 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9wm9\" (UniqueName: \"kubernetes.io/projected/ef6b7b00-8f47-4d05-92f0-b43367f92c12-kube-api-access-l9wm9\") pod \"heat-cfnapi-6d4656b7b7-t62gs\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.803145 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-55bc76fdd7-jqgvd"] Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.830363 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.865648 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.865770 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-combined-ca-bundle\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.865843 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnnhb\" (UniqueName: \"kubernetes.io/projected/5712fffb-e277-4ebd-a130-fee6f21273dc-kube-api-access-hnnhb\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.865918 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data-custom\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.877036 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data-custom\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.883837 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.885924 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-combined-ca-bundle\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.931328 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnnhb\" (UniqueName: \"kubernetes.io/projected/5712fffb-e277-4ebd-a130-fee6f21273dc-kube-api-access-hnnhb\") pod \"heat-api-55bc76fdd7-jqgvd\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:51 crc kubenswrapper[4912]: I1203 00:49:51.990487 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.014659 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.131478 4912 generic.go:334] "Generic (PLEG): container finished" podID="caa68042-1653-4df5-9249-6263d3fcbd52" containerID="1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428" exitCode=0 Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.131819 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.133030 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerDied","Data":"1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428"} Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.133115 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"caa68042-1653-4df5-9249-6263d3fcbd52","Type":"ContainerDied","Data":"1bb5ea542a2cd85995a933b9671dde2cb205c71e99ab775e919d1979316bbcdb"} Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.133136 4912 scope.go:117] "RemoveContainer" containerID="b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.176498 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-run-httpd\") pod \"caa68042-1653-4df5-9249-6263d3fcbd52\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.176590 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-sg-core-conf-yaml\") pod \"caa68042-1653-4df5-9249-6263d3fcbd52\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.176649 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-config-data\") pod \"caa68042-1653-4df5-9249-6263d3fcbd52\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.176745 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46p4d\" (UniqueName: \"kubernetes.io/projected/caa68042-1653-4df5-9249-6263d3fcbd52-kube-api-access-46p4d\") pod \"caa68042-1653-4df5-9249-6263d3fcbd52\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.176791 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-scripts\") pod \"caa68042-1653-4df5-9249-6263d3fcbd52\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.176836 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-log-httpd\") pod \"caa68042-1653-4df5-9249-6263d3fcbd52\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.176937 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-combined-ca-bundle\") pod \"caa68042-1653-4df5-9249-6263d3fcbd52\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.183050 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "caa68042-1653-4df5-9249-6263d3fcbd52" (UID: "caa68042-1653-4df5-9249-6263d3fcbd52"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.184248 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "caa68042-1653-4df5-9249-6263d3fcbd52" (UID: "caa68042-1653-4df5-9249-6263d3fcbd52"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.202591 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-scripts" (OuterVolumeSpecName: "scripts") pod "caa68042-1653-4df5-9249-6263d3fcbd52" (UID: "caa68042-1653-4df5-9249-6263d3fcbd52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.204570 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa68042-1653-4df5-9249-6263d3fcbd52-kube-api-access-46p4d" (OuterVolumeSpecName: "kube-api-access-46p4d") pod "caa68042-1653-4df5-9249-6263d3fcbd52" (UID: "caa68042-1653-4df5-9249-6263d3fcbd52"). InnerVolumeSpecName "kube-api-access-46p4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.214636 4912 scope.go:117] "RemoveContainer" containerID="3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.255764 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "caa68042-1653-4df5-9249-6263d3fcbd52" (UID: "caa68042-1653-4df5-9249-6263d3fcbd52"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.261253 4912 scope.go:117] "RemoveContainer" containerID="1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.294666 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.294720 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.294739 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46p4d\" (UniqueName: \"kubernetes.io/projected/caa68042-1653-4df5-9249-6263d3fcbd52-kube-api-access-46p4d\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.294756 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.294771 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/caa68042-1653-4df5-9249-6263d3fcbd52-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.318089 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-78c76df587-7nwws"] Dec 03 00:49:52 crc kubenswrapper[4912]: W1203 00:49:52.373924 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4881ee51_d772_4954_8e62_29584cc7b3c8.slice/crio-75aa1ff0e847c2a0044814c742acc65f0a4f881516e5a266413e7892969462b7 WatchSource:0}: Error finding container 75aa1ff0e847c2a0044814c742acc65f0a4f881516e5a266413e7892969462b7: Status 404 returned error can't find the container with id 75aa1ff0e847c2a0044814c742acc65f0a4f881516e5a266413e7892969462b7 Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.408470 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "caa68042-1653-4df5-9249-6263d3fcbd52" (UID: "caa68042-1653-4df5-9249-6263d3fcbd52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.408801 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-combined-ca-bundle\") pod \"caa68042-1653-4df5-9249-6263d3fcbd52\" (UID: \"caa68042-1653-4df5-9249-6263d3fcbd52\") " Dec 03 00:49:52 crc kubenswrapper[4912]: W1203 00:49:52.408941 4912 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/caa68042-1653-4df5-9249-6263d3fcbd52/volumes/kubernetes.io~secret/combined-ca-bundle Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.408964 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "caa68042-1653-4df5-9249-6263d3fcbd52" (UID: "caa68042-1653-4df5-9249-6263d3fcbd52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.410302 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.439537 4912 scope.go:117] "RemoveContainer" containerID="1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.455562 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-config-data" (OuterVolumeSpecName: "config-data") pod "caa68042-1653-4df5-9249-6263d3fcbd52" (UID: "caa68042-1653-4df5-9249-6263d3fcbd52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.509780 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.512680 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/caa68042-1653-4df5-9249-6263d3fcbd52-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.573115 4912 scope.go:117] "RemoveContainer" containerID="b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc" Dec 03 00:49:52 crc kubenswrapper[4912]: E1203 00:49:52.578468 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc\": container with ID starting with b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc not found: ID does not exist" containerID="b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.578525 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc"} err="failed to get container status \"b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc\": rpc error: code = NotFound desc = could not find container \"b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc\": container with ID starting with b6d7cf00c3b7a2462541bceae358b0adc68e2981af2260ff4b5ee07270c3a9fc not found: ID does not exist" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.578556 4912 scope.go:117] "RemoveContainer" containerID="3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5" Dec 03 00:49:52 crc kubenswrapper[4912]: E1203 00:49:52.581611 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5\": container with ID starting with 3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5 not found: ID does not exist" containerID="3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.581685 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5"} err="failed to get container status \"3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5\": rpc error: code = NotFound desc = could not find container \"3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5\": container with ID starting with 3bc6466d4cfd361fae9a2c62a2c556575ab89511cbd1b4f0707e412d2f58e1f5 not found: ID does not exist" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.581737 4912 scope.go:117] "RemoveContainer" containerID="1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428" Dec 03 00:49:52 crc kubenswrapper[4912]: E1203 00:49:52.582135 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428\": container with ID starting with 1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428 not found: ID does not exist" containerID="1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.582181 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428"} err="failed to get container status \"1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428\": rpc error: code = NotFound desc = could not find container \"1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428\": container with ID starting with 1f0ea7f9f0b8bbf3eba2af143e298e1fe1ccecafc4700245ee267df06a169428 not found: ID does not exist" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.582198 4912 scope.go:117] "RemoveContainer" containerID="1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1" Dec 03 00:49:52 crc kubenswrapper[4912]: E1203 00:49:52.582460 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1\": container with ID starting with 1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1 not found: ID does not exist" containerID="1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.582488 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1"} err="failed to get container status \"1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1\": rpc error: code = NotFound desc = could not find container \"1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1\": container with ID starting with 1f74c121dabbc66a9669d11ce6e3e901ccf754e58492d8f8c0e577f030ef4bb1 not found: ID does not exist" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.803637 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.816276 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.880546 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:52 crc kubenswrapper[4912]: E1203 00:49:52.881256 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="ceilometer-notification-agent" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.881278 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="ceilometer-notification-agent" Dec 03 00:49:52 crc kubenswrapper[4912]: E1203 00:49:52.881296 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="sg-core" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.881305 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="sg-core" Dec 03 00:49:52 crc kubenswrapper[4912]: E1203 00:49:52.881319 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="proxy-httpd" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.881325 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="proxy-httpd" Dec 03 00:49:52 crc kubenswrapper[4912]: E1203 00:49:52.881366 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="ceilometer-central-agent" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.881372 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="ceilometer-central-agent" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.881622 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="proxy-httpd" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.881647 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="ceilometer-notification-agent" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.881657 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="sg-core" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.881670 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" containerName="ceilometer-central-agent" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.883966 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.886895 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.887210 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.902914 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.928654 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-run-httpd\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.928728 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbdfj\" (UniqueName: \"kubernetes.io/projected/47d51262-7f59-4001-987b-f17c3b9a13dc-kube-api-access-sbdfj\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.928921 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.928981 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-log-httpd\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.929010 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-config-data\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.929047 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-scripts\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:52 crc kubenswrapper[4912]: I1203 00:49:52.929078 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.031576 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-log-httpd\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.031633 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-config-data\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.031683 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-scripts\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.031715 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.031783 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-run-httpd\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.031812 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbdfj\" (UniqueName: \"kubernetes.io/projected/47d51262-7f59-4001-987b-f17c3b9a13dc-kube-api-access-sbdfj\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.031963 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.032346 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-log-httpd\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.032652 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-run-httpd\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.040233 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.043293 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-scripts\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.045564 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-config-data\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.056556 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbdfj\" (UniqueName: \"kubernetes.io/projected/47d51262-7f59-4001-987b-f17c3b9a13dc-kube-api-access-sbdfj\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.068041 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.221416 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.235932 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-78c76df587-7nwws" event={"ID":"4881ee51-d772-4954-8e62-29584cc7b3c8","Type":"ContainerStarted","Data":"2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5"} Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.236009 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-78c76df587-7nwws" event={"ID":"4881ee51-d772-4954-8e62-29584cc7b3c8","Type":"ContainerStarted","Data":"75aa1ff0e847c2a0044814c742acc65f0a4f881516e5a266413e7892969462b7"} Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.238038 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.299044 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-78c76df587-7nwws" podStartSLOduration=3.299014783 podStartE2EDuration="3.299014783s" podCreationTimestamp="2025-12-03 00:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:53.298788106 +0000 UTC m=+1578.940808676" watchObservedRunningTime="2025-12-03 00:49:53.299014783 +0000 UTC m=+1578.941035343" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.485643 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.490153 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.512238 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.513055 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.581120 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn4xm\" (UniqueName: \"kubernetes.io/projected/4cea1b95-ee14-4157-80f0-8785715d76fb-kube-api-access-gn4xm\") pod \"4cea1b95-ee14-4157-80f0-8785715d76fb\" (UID: \"4cea1b95-ee14-4157-80f0-8785715d76fb\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.581210 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs6lb\" (UniqueName: \"kubernetes.io/projected/5e0fc17b-4a05-4eef-87d9-1269183df954-kube-api-access-bs6lb\") pod \"5e0fc17b-4a05-4eef-87d9-1269183df954\" (UID: \"5e0fc17b-4a05-4eef-87d9-1269183df954\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.581318 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-operator-scripts\") pod \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\" (UID: \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.581344 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxs8t\" (UniqueName: \"kubernetes.io/projected/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-kube-api-access-vxs8t\") pod \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\" (UID: \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.581369 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cea1b95-ee14-4157-80f0-8785715d76fb-operator-scripts\") pod \"4cea1b95-ee14-4157-80f0-8785715d76fb\" (UID: \"4cea1b95-ee14-4157-80f0-8785715d76fb\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.581453 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d27d\" (UniqueName: \"kubernetes.io/projected/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-kube-api-access-4d27d\") pod \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\" (UID: \"8f8ba108-0d7f-4399-a95f-aaf6b65b3499\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.581501 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e0fc17b-4a05-4eef-87d9-1269183df954-operator-scripts\") pod \"5e0fc17b-4a05-4eef-87d9-1269183df954\" (UID: \"5e0fc17b-4a05-4eef-87d9-1269183df954\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.581541 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-operator-scripts\") pod \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\" (UID: \"3d9ac0c8-42e7-4d85-be36-fd702a7c8787\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.582964 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3d9ac0c8-42e7-4d85-be36-fd702a7c8787" (UID: "3d9ac0c8-42e7-4d85-be36-fd702a7c8787"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.583403 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e0fc17b-4a05-4eef-87d9-1269183df954-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5e0fc17b-4a05-4eef-87d9-1269183df954" (UID: "5e0fc17b-4a05-4eef-87d9-1269183df954"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.585930 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8f8ba108-0d7f-4399-a95f-aaf6b65b3499" (UID: "8f8ba108-0d7f-4399-a95f-aaf6b65b3499"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.586929 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cea1b95-ee14-4157-80f0-8785715d76fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4cea1b95-ee14-4157-80f0-8785715d76fb" (UID: "4cea1b95-ee14-4157-80f0-8785715d76fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.602799 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-kube-api-access-4d27d" (OuterVolumeSpecName: "kube-api-access-4d27d") pod "8f8ba108-0d7f-4399-a95f-aaf6b65b3499" (UID: "8f8ba108-0d7f-4399-a95f-aaf6b65b3499"). InnerVolumeSpecName "kube-api-access-4d27d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.611983 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-kube-api-access-vxs8t" (OuterVolumeSpecName: "kube-api-access-vxs8t") pod "3d9ac0c8-42e7-4d85-be36-fd702a7c8787" (UID: "3d9ac0c8-42e7-4d85-be36-fd702a7c8787"). InnerVolumeSpecName "kube-api-access-vxs8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.612060 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cea1b95-ee14-4157-80f0-8785715d76fb-kube-api-access-gn4xm" (OuterVolumeSpecName: "kube-api-access-gn4xm") pod "4cea1b95-ee14-4157-80f0-8785715d76fb" (UID: "4cea1b95-ee14-4157-80f0-8785715d76fb"). InnerVolumeSpecName "kube-api-access-gn4xm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.613302 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.619521 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.626148 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e0fc17b-4a05-4eef-87d9-1269183df954-kube-api-access-bs6lb" (OuterVolumeSpecName: "kube-api-access-bs6lb") pod "5e0fc17b-4a05-4eef-87d9-1269183df954" (UID: "5e0fc17b-4a05-4eef-87d9-1269183df954"). InnerVolumeSpecName "kube-api-access-bs6lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.688700 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncbtk\" (UniqueName: \"kubernetes.io/projected/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-kube-api-access-ncbtk\") pod \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\" (UID: \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.688886 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-operator-scripts\") pod \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\" (UID: \"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.689017 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-operator-scripts\") pod \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\" (UID: \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.689057 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgrm6\" (UniqueName: \"kubernetes.io/projected/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-kube-api-access-xgrm6\") pod \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\" (UID: \"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72\") " Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.690092 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9" (UID: "4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.695002 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae1a6a50-b1c5-43b9-994c-7d3072dd8f72" (UID: "ae1a6a50-b1c5-43b9-994c-7d3072dd8f72"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.697770 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.697827 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxs8t\" (UniqueName: \"kubernetes.io/projected/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-kube-api-access-vxs8t\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.697848 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cea1b95-ee14-4157-80f0-8785715d76fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.697862 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d27d\" (UniqueName: \"kubernetes.io/projected/8f8ba108-0d7f-4399-a95f-aaf6b65b3499-kube-api-access-4d27d\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.697873 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.697885 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5e0fc17b-4a05-4eef-87d9-1269183df954-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.697902 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d9ac0c8-42e7-4d85-be36-fd702a7c8787-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.697913 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn4xm\" (UniqueName: \"kubernetes.io/projected/4cea1b95-ee14-4157-80f0-8785715d76fb-kube-api-access-gn4xm\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.697922 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs6lb\" (UniqueName: \"kubernetes.io/projected/5e0fc17b-4a05-4eef-87d9-1269183df954-kube-api-access-bs6lb\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.702805 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-kube-api-access-ncbtk" (OuterVolumeSpecName: "kube-api-access-ncbtk") pod "4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9" (UID: "4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9"). InnerVolumeSpecName "kube-api-access-ncbtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.737660 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-kube-api-access-xgrm6" (OuterVolumeSpecName: "kube-api-access-xgrm6") pod "ae1a6a50-b1c5-43b9-994c-7d3072dd8f72" (UID: "ae1a6a50-b1c5-43b9-994c-7d3072dd8f72"). InnerVolumeSpecName "kube-api-access-xgrm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.800092 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncbtk\" (UniqueName: \"kubernetes.io/projected/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-kube-api-access-ncbtk\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.800138 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.800148 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgrm6\" (UniqueName: \"kubernetes.io/projected/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72-kube-api-access-xgrm6\") on node \"crc\" DevicePath \"\"" Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.841960 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6d4656b7b7-t62gs"] Dec 03 00:49:53 crc kubenswrapper[4912]: I1203 00:49:53.879967 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-g7nr8"] Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.050571 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-55bc76fdd7-jqgvd"] Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.293896 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.303998 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fqvcx" event={"ID":"8f8ba108-0d7f-4399-a95f-aaf6b65b3499","Type":"ContainerDied","Data":"6ee6a2720f17577561427c70e6e6e7cef36df3e6a2eea46e12953f11cef0ff9f"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.304048 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ee6a2720f17577561427c70e6e6e7cef36df3e6a2eea46e12953f11cef0ff9f" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.304170 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fqvcx" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.311924 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4b7wk" event={"ID":"5e0fc17b-4a05-4eef-87d9-1269183df954","Type":"ContainerDied","Data":"338b4f9ad38150948bd08761d1218ac1fbc5246e68644371441037bcb9b402e9"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.311975 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="338b4f9ad38150948bd08761d1218ac1fbc5246e68644371441037bcb9b402e9" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.312079 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4b7wk" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.317742 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" event={"ID":"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02","Type":"ContainerStarted","Data":"71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.317816 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" event={"ID":"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02","Type":"ContainerStarted","Data":"0bba508da0571926191f604728f757b5b98f2dd18da87dd797dfbf15e6c76dd9"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.329170 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-vk882" event={"ID":"4cea1b95-ee14-4157-80f0-8785715d76fb","Type":"ContainerDied","Data":"a015f44208388f30316e6cffc96be4e6fa46ff7ea1559617ad575f6e7d07244e"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.329215 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a015f44208388f30316e6cffc96be4e6fa46ff7ea1559617ad575f6e7d07244e" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.329311 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-vk882" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.339007 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-24b5-account-create-update-4glzv" event={"ID":"4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9","Type":"ContainerDied","Data":"fee9728e66ce8ecbcdd92c7131260f22a94403c713f3f202285e76a8337318cb"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.339211 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fee9728e66ce8ecbcdd92c7131260f22a94403c713f3f202285e76a8337318cb" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.339386 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-24b5-account-create-update-4glzv" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.351942 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.351967 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-81fe-account-create-update-8dcm2" event={"ID":"3d9ac0c8-42e7-4d85-be36-fd702a7c8787","Type":"ContainerDied","Data":"cc4501acf377a3d00632a15bfb1de5cf8d369be6dcb4f3bd6f9c19d223cdf7fa"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.353693 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc4501acf377a3d00632a15bfb1de5cf8d369be6dcb4f3bd6f9c19d223cdf7fa" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.380703 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-55bc76fdd7-jqgvd" event={"ID":"5712fffb-e277-4ebd-a130-fee6f21273dc","Type":"ContainerStarted","Data":"d9eb7767fbf79623eee66ce885f8544cd1398163a2aa9b424a131048328b9e2f"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.384642 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-9697-account-create-update-chh6n" event={"ID":"ae1a6a50-b1c5-43b9-994c-7d3072dd8f72","Type":"ContainerDied","Data":"bfafce7c2451377b8deb76263144d935b633fda987c7bce313f0e59fabbf57d9"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.385767 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfafce7c2451377b8deb76263144d935b633fda987c7bce313f0e59fabbf57d9" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.385879 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-9697-account-create-update-chh6n" Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.396022 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" event={"ID":"ef6b7b00-8f47-4d05-92f0-b43367f92c12","Type":"ContainerStarted","Data":"92a1770782fbf4f9c7c17e58866f4af5d740142f8296670e7c117c4d9c004a41"} Dec 03 00:49:54 crc kubenswrapper[4912]: I1203 00:49:54.614886 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caa68042-1653-4df5-9249-6263d3fcbd52" path="/var/lib/kubelet/pods/caa68042-1653-4df5-9249-6263d3fcbd52/volumes" Dec 03 00:49:55 crc kubenswrapper[4912]: I1203 00:49:55.417889 4912 generic.go:334] "Generic (PLEG): container finished" podID="dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" containerID="71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f" exitCode=0 Dec 03 00:49:55 crc kubenswrapper[4912]: I1203 00:49:55.418258 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" event={"ID":"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02","Type":"ContainerDied","Data":"71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f"} Dec 03 00:49:55 crc kubenswrapper[4912]: I1203 00:49:55.418872 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" event={"ID":"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02","Type":"ContainerStarted","Data":"b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea"} Dec 03 00:49:55 crc kubenswrapper[4912]: I1203 00:49:55.418908 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:49:55 crc kubenswrapper[4912]: I1203 00:49:55.426593 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerStarted","Data":"c51247a128514615bf0495e1e754f19ebe2d6a62c0154b6b806fc34eb3ff281c"} Dec 03 00:49:55 crc kubenswrapper[4912]: I1203 00:49:55.426640 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerStarted","Data":"55af0a981407746afd06137f48b16e76f720bdabd508bb06288627038a7271b0"} Dec 03 00:49:55 crc kubenswrapper[4912]: I1203 00:49:55.464514 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" podStartSLOduration=4.464491814 podStartE2EDuration="4.464491814s" podCreationTimestamp="2025-12-03 00:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:49:55.450740197 +0000 UTC m=+1581.092760747" watchObservedRunningTime="2025-12-03 00:49:55.464491814 +0000 UTC m=+1581.106512384" Dec 03 00:49:56 crc kubenswrapper[4912]: I1203 00:49:56.726930 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:56 crc kubenswrapper[4912]: I1203 00:49:56.728741 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7ddd66b7bc-vwpcg" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.499610 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" event={"ID":"ef6b7b00-8f47-4d05-92f0-b43367f92c12","Type":"ContainerStarted","Data":"7cbac6e5060818ac2503d5b330e04bb6e951432dda9df14f01b03998cf9aabfe"} Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.500632 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.505026 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerStarted","Data":"87c8beda356bc1e9b4bfa123fc88bec6bcd22bdff115f65e76e91d068738533c"} Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.510460 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-55bc76fdd7-jqgvd" event={"ID":"5712fffb-e277-4ebd-a130-fee6f21273dc","Type":"ContainerStarted","Data":"80efefa6f5b5386935076ff537b2269625ea692dc72a2686a831a0a282f48a08"} Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.511636 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.522520 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" podStartSLOduration=4.045780822 podStartE2EDuration="7.522500383s" podCreationTimestamp="2025-12-03 00:49:51 +0000 UTC" firstStartedPulling="2025-12-03 00:49:53.898676817 +0000 UTC m=+1579.540697377" lastFinishedPulling="2025-12-03 00:49:57.375396378 +0000 UTC m=+1583.017416938" observedRunningTime="2025-12-03 00:49:58.520478289 +0000 UTC m=+1584.162498849" watchObservedRunningTime="2025-12-03 00:49:58.522500383 +0000 UTC m=+1584.164520953" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.604197 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-55bc76fdd7-jqgvd" podStartSLOduration=4.290576659 podStartE2EDuration="7.604159924s" podCreationTimestamp="2025-12-03 00:49:51 +0000 UTC" firstStartedPulling="2025-12-03 00:49:54.057305003 +0000 UTC m=+1579.699325563" lastFinishedPulling="2025-12-03 00:49:57.370888268 +0000 UTC m=+1583.012908828" observedRunningTime="2025-12-03 00:49:58.553339476 +0000 UTC m=+1584.195360036" watchObservedRunningTime="2025-12-03 00:49:58.604159924 +0000 UTC m=+1584.246180484" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.990846 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wpgr9"] Dec 03 00:49:58 crc kubenswrapper[4912]: E1203 00:49:58.991395 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8ba108-0d7f-4399-a95f-aaf6b65b3499" containerName="mariadb-database-create" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991414 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8ba108-0d7f-4399-a95f-aaf6b65b3499" containerName="mariadb-database-create" Dec 03 00:49:58 crc kubenswrapper[4912]: E1203 00:49:58.991447 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1a6a50-b1c5-43b9-994c-7d3072dd8f72" containerName="mariadb-account-create-update" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991454 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1a6a50-b1c5-43b9-994c-7d3072dd8f72" containerName="mariadb-account-create-update" Dec 03 00:49:58 crc kubenswrapper[4912]: E1203 00:49:58.991476 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e0fc17b-4a05-4eef-87d9-1269183df954" containerName="mariadb-database-create" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991482 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e0fc17b-4a05-4eef-87d9-1269183df954" containerName="mariadb-database-create" Dec 03 00:49:58 crc kubenswrapper[4912]: E1203 00:49:58.991508 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cea1b95-ee14-4157-80f0-8785715d76fb" containerName="mariadb-database-create" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991515 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cea1b95-ee14-4157-80f0-8785715d76fb" containerName="mariadb-database-create" Dec 03 00:49:58 crc kubenswrapper[4912]: E1203 00:49:58.991545 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9ac0c8-42e7-4d85-be36-fd702a7c8787" containerName="mariadb-account-create-update" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991551 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9ac0c8-42e7-4d85-be36-fd702a7c8787" containerName="mariadb-account-create-update" Dec 03 00:49:58 crc kubenswrapper[4912]: E1203 00:49:58.991566 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9" containerName="mariadb-account-create-update" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991575 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9" containerName="mariadb-account-create-update" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991779 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9" containerName="mariadb-account-create-update" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991788 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae1a6a50-b1c5-43b9-994c-7d3072dd8f72" containerName="mariadb-account-create-update" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991798 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9ac0c8-42e7-4d85-be36-fd702a7c8787" containerName="mariadb-account-create-update" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991814 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e0fc17b-4a05-4eef-87d9-1269183df954" containerName="mariadb-database-create" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991824 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cea1b95-ee14-4157-80f0-8785715d76fb" containerName="mariadb-database-create" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.991840 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8ba108-0d7f-4399-a95f-aaf6b65b3499" containerName="mariadb-database-create" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.992678 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.995638 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-4frvc" Dec 03 00:49:58 crc kubenswrapper[4912]: I1203 00:49:58.995840 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.000006 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.020809 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wpgr9"] Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.075804 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgvwj\" (UniqueName: \"kubernetes.io/projected/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-kube-api-access-zgvwj\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.076290 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-config-data\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.076314 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.076346 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-scripts\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.179126 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgvwj\" (UniqueName: \"kubernetes.io/projected/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-kube-api-access-zgvwj\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.179356 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-config-data\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.179400 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.179495 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-scripts\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.189297 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-config-data\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.189654 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-scripts\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.192115 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.201609 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgvwj\" (UniqueName: \"kubernetes.io/projected/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-kube-api-access-zgvwj\") pod \"nova-cell0-conductor-db-sync-wpgr9\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.313773 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.563485 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerStarted","Data":"83fcf4cdd73539bf838749b3754c62aa67a8e5cf7dcea3e10baebdb6bb591d9f"} Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.569495 4912 generic.go:334] "Generic (PLEG): container finished" podID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerID="eb2afd092785dbdfb95cd0b170a4e990e35a2115166ac5a87cf5a04eb82c43a5" exitCode=137 Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.571225 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22763c05-583a-4681-bbf4-882fd2a8d7ed","Type":"ContainerDied","Data":"eb2afd092785dbdfb95cd0b170a4e990e35a2115166ac5a87cf5a04eb82c43a5"} Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.737006 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5fc6c75d4d-bwk2j"] Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.740260 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.780739 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5fc6c75d4d-bwk2j"] Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.814710 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-694c75fc98-gbk8k"] Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.847526 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6b486667b8-shqw9"] Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.848976 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.849714 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.856528 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-694c75fc98-gbk8k"] Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.894935 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6b486667b8-shqw9"] Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.907303 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxzqm\" (UniqueName: \"kubernetes.io/projected/74900d46-53f3-4cd5-aa56-9509c80b0041-kube-api-access-jxzqm\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.909233 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.909508 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data-custom\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:49:59 crc kubenswrapper[4912]: I1203 00:49:59.909951 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-combined-ca-bundle\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.012846 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.012971 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.012997 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dlp8\" (UniqueName: \"kubernetes.io/projected/04d938cc-1945-46d6-bd97-4c78397f713a-kube-api-access-6dlp8\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.013119 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-combined-ca-bundle\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.013225 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.013297 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data-custom\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.013317 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data-custom\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.014006 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-combined-ca-bundle\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.014234 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-combined-ca-bundle\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.014545 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data-custom\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.014612 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxzqm\" (UniqueName: \"kubernetes.io/projected/74900d46-53f3-4cd5-aa56-9509c80b0041-kube-api-access-jxzqm\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.014656 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtfw4\" (UniqueName: \"kubernetes.io/projected/10282fe1-8cff-49d3-8d01-78c23a691eac-kube-api-access-xtfw4\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.021201 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.023177 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data-custom\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.024275 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-combined-ca-bundle\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.036509 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxzqm\" (UniqueName: \"kubernetes.io/projected/74900d46-53f3-4cd5-aa56-9509c80b0041-kube-api-access-jxzqm\") pod \"heat-engine-5fc6c75d4d-bwk2j\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.117196 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.117354 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dlp8\" (UniqueName: \"kubernetes.io/projected/04d938cc-1945-46d6-bd97-4c78397f713a-kube-api-access-6dlp8\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.117385 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-combined-ca-bundle\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.117486 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.117533 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data-custom\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.117612 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-combined-ca-bundle\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.117661 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data-custom\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.117711 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtfw4\" (UniqueName: \"kubernetes.io/projected/10282fe1-8cff-49d3-8d01-78c23a691eac-kube-api-access-xtfw4\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.137333 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data-custom\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.138156 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.140321 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-combined-ca-bundle\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.140783 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-combined-ca-bundle\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.142203 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dlp8\" (UniqueName: \"kubernetes.io/projected/04d938cc-1945-46d6-bd97-4c78397f713a-kube-api-access-6dlp8\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.143043 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.144248 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wpgr9"] Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.147403 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtfw4\" (UniqueName: \"kubernetes.io/projected/10282fe1-8cff-49d3-8d01-78c23a691eac-kube-api-access-xtfw4\") pod \"heat-api-6b486667b8-shqw9\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.153872 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data-custom\") pod \"heat-cfnapi-694c75fc98-gbk8k\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.163080 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.182540 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.202974 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.280247 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.432948 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data-custom\") pod \"22763c05-583a-4681-bbf4-882fd2a8d7ed\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.433170 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4488v\" (UniqueName: \"kubernetes.io/projected/22763c05-583a-4681-bbf4-882fd2a8d7ed-kube-api-access-4488v\") pod \"22763c05-583a-4681-bbf4-882fd2a8d7ed\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.433258 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-scripts\") pod \"22763c05-583a-4681-bbf4-882fd2a8d7ed\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.433343 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22763c05-583a-4681-bbf4-882fd2a8d7ed-etc-machine-id\") pod \"22763c05-583a-4681-bbf4-882fd2a8d7ed\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.433458 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-combined-ca-bundle\") pod \"22763c05-583a-4681-bbf4-882fd2a8d7ed\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.433486 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data\") pod \"22763c05-583a-4681-bbf4-882fd2a8d7ed\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.433549 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22763c05-583a-4681-bbf4-882fd2a8d7ed-logs\") pod \"22763c05-583a-4681-bbf4-882fd2a8d7ed\" (UID: \"22763c05-583a-4681-bbf4-882fd2a8d7ed\") " Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.434859 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22763c05-583a-4681-bbf4-882fd2a8d7ed-logs" (OuterVolumeSpecName: "logs") pod "22763c05-583a-4681-bbf4-882fd2a8d7ed" (UID: "22763c05-583a-4681-bbf4-882fd2a8d7ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.434912 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22763c05-583a-4681-bbf4-882fd2a8d7ed-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "22763c05-583a-4681-bbf4-882fd2a8d7ed" (UID: "22763c05-583a-4681-bbf4-882fd2a8d7ed"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.442167 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22763c05-583a-4681-bbf4-882fd2a8d7ed-kube-api-access-4488v" (OuterVolumeSpecName: "kube-api-access-4488v") pod "22763c05-583a-4681-bbf4-882fd2a8d7ed" (UID: "22763c05-583a-4681-bbf4-882fd2a8d7ed"). InnerVolumeSpecName "kube-api-access-4488v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.448260 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "22763c05-583a-4681-bbf4-882fd2a8d7ed" (UID: "22763c05-583a-4681-bbf4-882fd2a8d7ed"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.453755 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-scripts" (OuterVolumeSpecName: "scripts") pod "22763c05-583a-4681-bbf4-882fd2a8d7ed" (UID: "22763c05-583a-4681-bbf4-882fd2a8d7ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.527203 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22763c05-583a-4681-bbf4-882fd2a8d7ed" (UID: "22763c05-583a-4681-bbf4-882fd2a8d7ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.536387 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.536652 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4488v\" (UniqueName: \"kubernetes.io/projected/22763c05-583a-4681-bbf4-882fd2a8d7ed-kube-api-access-4488v\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.536743 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.536823 4912 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22763c05-583a-4681-bbf4-882fd2a8d7ed-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.536918 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.537011 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22763c05-583a-4681-bbf4-882fd2a8d7ed-logs\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.574579 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data" (OuterVolumeSpecName: "config-data") pod "22763c05-583a-4681-bbf4-882fd2a8d7ed" (UID: "22763c05-583a-4681-bbf4-882fd2a8d7ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.598963 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.599285 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"22763c05-583a-4681-bbf4-882fd2a8d7ed","Type":"ContainerDied","Data":"0caf1d4d8b9be3f1786fb42971254e26ad6101685c058f7cb27639349f08d83c"} Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.599455 4912 scope.go:117] "RemoveContainer" containerID="eb2afd092785dbdfb95cd0b170a4e990e35a2115166ac5a87cf5a04eb82c43a5" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.612073 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" event={"ID":"aeac5989-b199-4fcf-8d5e-c49ac6062bfd","Type":"ContainerStarted","Data":"c325707086b379052608ca92ce1fb4bc904c5db2c7e10f44386aaaa09194b05c"} Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.636243 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerStarted","Data":"c4f4514f1bd3168e24977a9dd97bf97b44cb2d4206b72a3630d9bce3c221698a"} Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.636328 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.640613 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22763c05-583a-4681-bbf4-882fd2a8d7ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.656261 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.707663 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.716538 4912 scope.go:117] "RemoveContainer" containerID="0c4c3865b6e4d25992684d34e8ec5c5a1579024e07c7418c6c44b071c311be45" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.766799 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 00:50:00 crc kubenswrapper[4912]: E1203 00:50:00.768419 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerName="cinder-api" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.768458 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerName="cinder-api" Dec 03 00:50:00 crc kubenswrapper[4912]: E1203 00:50:00.768484 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerName="cinder-api-log" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.768494 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerName="cinder-api-log" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.774655 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerName="cinder-api-log" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.774754 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="22763c05-583a-4681-bbf4-882fd2a8d7ed" containerName="cinder-api" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.781847 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.789840 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.790208 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.790359 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.814203 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.080382855 podStartE2EDuration="8.814171955s" podCreationTimestamp="2025-12-03 00:49:52 +0000 UTC" firstStartedPulling="2025-12-03 00:49:54.362888744 +0000 UTC m=+1580.004909304" lastFinishedPulling="2025-12-03 00:50:00.096677824 +0000 UTC m=+1585.738698404" observedRunningTime="2025-12-03 00:50:00.670069617 +0000 UTC m=+1586.312090187" watchObservedRunningTime="2025-12-03 00:50:00.814171955 +0000 UTC m=+1586.456192515" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.899283 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.916696 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64841d14-78b5-4737-90ec-b812f2994397-logs\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.916808 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-public-tls-certs\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.916866 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crmpd\" (UniqueName: \"kubernetes.io/projected/64841d14-78b5-4737-90ec-b812f2994397-kube-api-access-crmpd\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.916894 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64841d14-78b5-4737-90ec-b812f2994397-etc-machine-id\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.917118 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-scripts\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.917325 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-config-data\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.917611 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-config-data-custom\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.918153 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:00 crc kubenswrapper[4912]: I1203 00:50:00.918613 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.021873 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-config-data\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.027632 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-config-data-custom\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.028550 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.029077 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.029176 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64841d14-78b5-4737-90ec-b812f2994397-logs\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.029325 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-public-tls-certs\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.030927 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-config-data\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.032942 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64841d14-78b5-4737-90ec-b812f2994397-logs\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.034523 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crmpd\" (UniqueName: \"kubernetes.io/projected/64841d14-78b5-4737-90ec-b812f2994397-kube-api-access-crmpd\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.034606 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64841d14-78b5-4737-90ec-b812f2994397-etc-machine-id\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.034669 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-scripts\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.035134 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64841d14-78b5-4737-90ec-b812f2994397-etc-machine-id\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.048012 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-scripts\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.038139 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-694c75fc98-gbk8k"] Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.050695 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-config-data-custom\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.051222 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-public-tls-certs\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.065641 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crmpd\" (UniqueName: \"kubernetes.io/projected/64841d14-78b5-4737-90ec-b812f2994397-kube-api-access-crmpd\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.081915 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: W1203 00:50:01.085376 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74900d46_53f3_4cd5_aa56_9509c80b0041.slice/crio-c51c2285879b2a463630686ee09f4c9518c07cf4e564c88f69a394ee817729f0 WatchSource:0}: Error finding container c51c2285879b2a463630686ee09f4c9518c07cf4e564c88f69a394ee817729f0: Status 404 returned error can't find the container with id c51c2285879b2a463630686ee09f4c9518c07cf4e564c88f69a394ee817729f0 Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.086386 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5fc6c75d4d-bwk2j"] Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.090960 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64841d14-78b5-4737-90ec-b812f2994397-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"64841d14-78b5-4737-90ec-b812f2994397\") " pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.180162 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.210425 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6b486667b8-shqw9"] Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.683954 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b486667b8-shqw9" event={"ID":"10282fe1-8cff-49d3-8d01-78c23a691eac","Type":"ContainerStarted","Data":"dd782341efa3233d4466a7230a183085b4b829ef804285c581314db749cd4b75"} Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.684924 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b486667b8-shqw9" event={"ID":"10282fe1-8cff-49d3-8d01-78c23a691eac","Type":"ContainerStarted","Data":"61d9ee2ae71b22b31f025a67a67a908974d3505f5ac2a1b0a5e46f5977c84a16"} Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.685017 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.691732 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" event={"ID":"74900d46-53f3-4cd5-aa56-9509c80b0041","Type":"ContainerStarted","Data":"9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a"} Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.691807 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" event={"ID":"74900d46-53f3-4cd5-aa56-9509c80b0041","Type":"ContainerStarted","Data":"c51c2285879b2a463630686ee09f4c9518c07cf4e564c88f69a394ee817729f0"} Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.691833 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.696515 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" event={"ID":"04d938cc-1945-46d6-bd97-4c78397f713a","Type":"ContainerStarted","Data":"10d7e0516e3487c09d7926fdae77e88822b063fe9b7bc934a55a48cd5cae63e0"} Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.696587 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" event={"ID":"04d938cc-1945-46d6-bd97-4c78397f713a","Type":"ContainerStarted","Data":"2343dc29e4732fade042928207d1638357ecb48b88af44db995404019f245238"} Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.716776 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6b486667b8-shqw9" podStartSLOduration=2.71673694 podStartE2EDuration="2.71673694s" podCreationTimestamp="2025-12-03 00:49:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:50:01.707882013 +0000 UTC m=+1587.349902573" watchObservedRunningTime="2025-12-03 00:50:01.71673694 +0000 UTC m=+1587.358757500" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.716884 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.751498 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" podStartSLOduration=2.751468787 podStartE2EDuration="2.751468787s" podCreationTimestamp="2025-12-03 00:49:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:50:01.730152247 +0000 UTC m=+1587.372172807" watchObservedRunningTime="2025-12-03 00:50:01.751468787 +0000 UTC m=+1587.393489347" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.773262 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" podStartSLOduration=2.773235598 podStartE2EDuration="2.773235598s" podCreationTimestamp="2025-12-03 00:49:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:50:01.752475904 +0000 UTC m=+1587.394496484" watchObservedRunningTime="2025-12-03 00:50:01.773235598 +0000 UTC m=+1587.415256158" Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.868573 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.885039 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-zmrhq"] Dec 03 00:50:01 crc kubenswrapper[4912]: I1203 00:50:01.885599 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" containerName="dnsmasq-dns" containerID="cri-o://814304fa1d3f48577aeb5a367cdc99bdd58f1b535c6730312645d059b030a7ec" gracePeriod=10 Dec 03 00:50:01 crc kubenswrapper[4912]: W1203 00:50:01.915598 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64841d14_78b5_4737_90ec_b812f2994397.slice/crio-87838e7848e9dcc31c6660eb8915b33762ef90f46dd0b28e53c4fb55ca1244be WatchSource:0}: Error finding container 87838e7848e9dcc31c6660eb8915b33762ef90f46dd0b28e53c4fb55ca1244be: Status 404 returned error can't find the container with id 87838e7848e9dcc31c6660eb8915b33762ef90f46dd0b28e53c4fb55ca1244be Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.571858 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:50:02 crc kubenswrapper[4912]: E1203 00:50:02.572191 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.598678 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22763c05-583a-4681-bbf4-882fd2a8d7ed" path="/var/lib/kubelet/pods/22763c05-583a-4681-bbf4-882fd2a8d7ed/volumes" Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.718083 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"64841d14-78b5-4737-90ec-b812f2994397","Type":"ContainerStarted","Data":"87838e7848e9dcc31c6660eb8915b33762ef90f46dd0b28e53c4fb55ca1244be"} Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.722913 4912 generic.go:334] "Generic (PLEG): container finished" podID="04d938cc-1945-46d6-bd97-4c78397f713a" containerID="10d7e0516e3487c09d7926fdae77e88822b063fe9b7bc934a55a48cd5cae63e0" exitCode=1 Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.723400 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" event={"ID":"04d938cc-1945-46d6-bd97-4c78397f713a","Type":"ContainerDied","Data":"10d7e0516e3487c09d7926fdae77e88822b063fe9b7bc934a55a48cd5cae63e0"} Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.724535 4912 scope.go:117] "RemoveContainer" containerID="10d7e0516e3487c09d7926fdae77e88822b063fe9b7bc934a55a48cd5cae63e0" Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.731779 4912 generic.go:334] "Generic (PLEG): container finished" podID="26f3f384-6499-4824-8c10-de6382e27d8b" containerID="814304fa1d3f48577aeb5a367cdc99bdd58f1b535c6730312645d059b030a7ec" exitCode=0 Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.731858 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" event={"ID":"26f3f384-6499-4824-8c10-de6382e27d8b","Type":"ContainerDied","Data":"814304fa1d3f48577aeb5a367cdc99bdd58f1b535c6730312645d059b030a7ec"} Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.734298 4912 generic.go:334] "Generic (PLEG): container finished" podID="10282fe1-8cff-49d3-8d01-78c23a691eac" containerID="dd782341efa3233d4466a7230a183085b4b829ef804285c581314db749cd4b75" exitCode=1 Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.751345 4912 scope.go:117] "RemoveContainer" containerID="dd782341efa3233d4466a7230a183085b4b829ef804285c581314db749cd4b75" Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.756766 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b486667b8-shqw9" event={"ID":"10282fe1-8cff-49d3-8d01-78c23a691eac","Type":"ContainerDied","Data":"dd782341efa3233d4466a7230a183085b4b829ef804285c581314db749cd4b75"} Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.965670 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-55bc76fdd7-jqgvd"] Dec 03 00:50:02 crc kubenswrapper[4912]: I1203 00:50:02.965956 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-55bc76fdd7-jqgvd" podUID="5712fffb-e277-4ebd-a130-fee6f21273dc" containerName="heat-api" containerID="cri-o://80efefa6f5b5386935076ff537b2269625ea692dc72a2686a831a0a282f48a08" gracePeriod=60 Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.019543 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6d4656b7b7-t62gs"] Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.019830 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" podUID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" containerName="heat-cfnapi" containerID="cri-o://7cbac6e5060818ac2503d5b330e04bb6e951432dda9df14f01b03998cf9aabfe" gracePeriod=60 Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.058903 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-69f9575dc5-s25p4"] Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.060645 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.064976 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.065768 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.068839 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" podUID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.200:8000/healthcheck\": EOF" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.081788 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69f9575dc5-s25p4"] Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.095955 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7b79dbdd58-qznp2"] Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.098231 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.101669 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7b79dbdd58-qznp2"] Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.105846 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.106083 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.146517 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-public-tls-certs\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.146576 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.146652 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-combined-ca-bundle\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.146680 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data-custom\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.146753 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-internal-tls-certs\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.146799 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q64n7\" (UniqueName: \"kubernetes.io/projected/c7638b15-f910-4a20-b770-c65ebdc8ac49-kube-api-access-q64n7\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.249753 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-internal-tls-certs\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.249815 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.249858 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q64n7\" (UniqueName: \"kubernetes.io/projected/c7638b15-f910-4a20-b770-c65ebdc8ac49-kube-api-access-q64n7\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.249916 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-internal-tls-certs\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.249965 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-public-tls-certs\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.249986 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-public-tls-certs\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.250008 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data-custom\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.250024 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.250075 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-combined-ca-bundle\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.250100 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqmwj\" (UniqueName: \"kubernetes.io/projected/d628e1f8-fd25-4bba-a85b-58f71eb64f58-kube-api-access-dqmwj\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.250126 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-combined-ca-bundle\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.250154 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data-custom\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.276227 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-internal-tls-certs\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.276538 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-public-tls-certs\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.276957 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-combined-ca-bundle\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.277059 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data-custom\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.278164 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.292877 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q64n7\" (UniqueName: \"kubernetes.io/projected/c7638b15-f910-4a20-b770-c65ebdc8ac49-kube-api-access-q64n7\") pod \"heat-api-69f9575dc5-s25p4\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.360075 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.360183 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-internal-tls-certs\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.360225 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-public-tls-certs\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.360250 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data-custom\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.360307 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-combined-ca-bundle\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.360334 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqmwj\" (UniqueName: \"kubernetes.io/projected/d628e1f8-fd25-4bba-a85b-58f71eb64f58-kube-api-access-dqmwj\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.366572 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.369331 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-combined-ca-bundle\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.369554 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-internal-tls-certs\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.374267 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-public-tls-certs\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.383584 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data-custom\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.387945 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqmwj\" (UniqueName: \"kubernetes.io/projected/d628e1f8-fd25-4bba-a85b-58f71eb64f58-kube-api-access-dqmwj\") pod \"heat-cfnapi-7b79dbdd58-qznp2\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.446494 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.476119 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.762778 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"64841d14-78b5-4737-90ec-b812f2994397","Type":"ContainerStarted","Data":"c7e6028cc55a729dee5fb73c5c3c79135cf9edf4ac59ada07dbff6b8b3b1eb2a"} Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.766728 4912 generic.go:334] "Generic (PLEG): container finished" podID="5712fffb-e277-4ebd-a130-fee6f21273dc" containerID="80efefa6f5b5386935076ff537b2269625ea692dc72a2686a831a0a282f48a08" exitCode=0 Dec 03 00:50:03 crc kubenswrapper[4912]: I1203 00:50:03.766768 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-55bc76fdd7-jqgvd" event={"ID":"5712fffb-e277-4ebd-a130-fee6f21273dc","Type":"ContainerDied","Data":"80efefa6f5b5386935076ff537b2269625ea692dc72a2686a831a0a282f48a08"} Dec 03 00:50:04 crc kubenswrapper[4912]: I1203 00:50:04.508783 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:04 crc kubenswrapper[4912]: I1203 00:50:04.509296 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="ceilometer-central-agent" containerID="cri-o://c51247a128514615bf0495e1e754f19ebe2d6a62c0154b6b806fc34eb3ff281c" gracePeriod=30 Dec 03 00:50:04 crc kubenswrapper[4912]: I1203 00:50:04.509619 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="ceilometer-notification-agent" containerID="cri-o://87c8beda356bc1e9b4bfa123fc88bec6bcd22bdff115f65e76e91d068738533c" gracePeriod=30 Dec 03 00:50:04 crc kubenswrapper[4912]: I1203 00:50:04.509608 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="proxy-httpd" containerID="cri-o://c4f4514f1bd3168e24977a9dd97bf97b44cb2d4206b72a3630d9bce3c221698a" gracePeriod=30 Dec 03 00:50:04 crc kubenswrapper[4912]: I1203 00:50:04.509680 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="sg-core" containerID="cri-o://83fcf4cdd73539bf838749b3754c62aa67a8e5cf7dcea3e10baebdb6bb591d9f" gracePeriod=30 Dec 03 00:50:04 crc kubenswrapper[4912]: I1203 00:50:04.788316 4912 generic.go:334] "Generic (PLEG): container finished" podID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerID="c4f4514f1bd3168e24977a9dd97bf97b44cb2d4206b72a3630d9bce3c221698a" exitCode=0 Dec 03 00:50:04 crc kubenswrapper[4912]: I1203 00:50:04.788357 4912 generic.go:334] "Generic (PLEG): container finished" podID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerID="83fcf4cdd73539bf838749b3754c62aa67a8e5cf7dcea3e10baebdb6bb591d9f" exitCode=2 Dec 03 00:50:04 crc kubenswrapper[4912]: I1203 00:50:04.788380 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerDied","Data":"c4f4514f1bd3168e24977a9dd97bf97b44cb2d4206b72a3630d9bce3c221698a"} Dec 03 00:50:04 crc kubenswrapper[4912]: I1203 00:50:04.788413 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerDied","Data":"83fcf4cdd73539bf838749b3754c62aa67a8e5cf7dcea3e10baebdb6bb591d9f"} Dec 03 00:50:05 crc kubenswrapper[4912]: I1203 00:50:05.183242 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:05 crc kubenswrapper[4912]: I1203 00:50:05.203303 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:05 crc kubenswrapper[4912]: I1203 00:50:05.203660 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:05 crc kubenswrapper[4912]: I1203 00:50:05.819400 4912 generic.go:334] "Generic (PLEG): container finished" podID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerID="87c8beda356bc1e9b4bfa123fc88bec6bcd22bdff115f65e76e91d068738533c" exitCode=0 Dec 03 00:50:05 crc kubenswrapper[4912]: I1203 00:50:05.819453 4912 generic.go:334] "Generic (PLEG): container finished" podID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerID="c51247a128514615bf0495e1e754f19ebe2d6a62c0154b6b806fc34eb3ff281c" exitCode=0 Dec 03 00:50:05 crc kubenswrapper[4912]: I1203 00:50:05.819477 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerDied","Data":"87c8beda356bc1e9b4bfa123fc88bec6bcd22bdff115f65e76e91d068738533c"} Dec 03 00:50:05 crc kubenswrapper[4912]: I1203 00:50:05.819508 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerDied","Data":"c51247a128514615bf0495e1e754f19ebe2d6a62c0154b6b806fc34eb3ff281c"} Dec 03 00:50:06 crc kubenswrapper[4912]: I1203 00:50:06.995009 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-55bc76fdd7-jqgvd" podUID="5712fffb-e277-4ebd-a130-fee6f21273dc" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.201:8004/healthcheck\": dial tcp 10.217.0.201:8004: connect: connection refused" Dec 03 00:50:08 crc kubenswrapper[4912]: I1203 00:50:08.454007 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" podUID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.200:8000/healthcheck\": read tcp 10.217.0.2:49010->10.217.0.200:8000: read: connection reset by peer" Dec 03 00:50:08 crc kubenswrapper[4912]: I1203 00:50:08.864746 4912 generic.go:334] "Generic (PLEG): container finished" podID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" containerID="7cbac6e5060818ac2503d5b330e04bb6e951432dda9df14f01b03998cf9aabfe" exitCode=0 Dec 03 00:50:08 crc kubenswrapper[4912]: I1203 00:50:08.864814 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" event={"ID":"ef6b7b00-8f47-4d05-92f0-b43367f92c12","Type":"ContainerDied","Data":"7cbac6e5060818ac2503d5b330e04bb6e951432dda9df14f01b03998cf9aabfe"} Dec 03 00:50:10 crc kubenswrapper[4912]: I1203 00:50:10.554090 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.185:5353: i/o timeout" Dec 03 00:50:11 crc kubenswrapper[4912]: I1203 00:50:11.408195 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:50:11 crc kubenswrapper[4912]: I1203 00:50:11.832360 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" podUID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.200:8000/healthcheck\": dial tcp 10.217.0.200:8000: connect: connection refused" Dec 03 00:50:11 crc kubenswrapper[4912]: I1203 00:50:11.992948 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-55bc76fdd7-jqgvd" podUID="5712fffb-e277-4ebd-a130-fee6f21273dc" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.201:8004/healthcheck\": dial tcp 10.217.0.201:8004: connect: connection refused" Dec 03 00:50:15 crc kubenswrapper[4912]: I1203 00:50:15.555675 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.185:5353: i/o timeout" Dec 03 00:50:15 crc kubenswrapper[4912]: I1203 00:50:15.572423 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:50:15 crc kubenswrapper[4912]: E1203 00:50:15.572822 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.254443 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:50:16 crc kubenswrapper[4912]: E1203 00:50:16.269206 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified" Dec 03 00:50:16 crc kubenswrapper[4912]: E1203 00:50:16.269373 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zgvwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-wpgr9_openstack(aeac5989-b199-4fcf-8d5e-c49ac6062bfd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:50:16 crc kubenswrapper[4912]: E1203 00:50:16.270731 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" podUID="aeac5989-b199-4fcf-8d5e-c49ac6062bfd" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.366540 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-config\") pod \"26f3f384-6499-4824-8c10-de6382e27d8b\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.367166 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-svc\") pod \"26f3f384-6499-4824-8c10-de6382e27d8b\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.367238 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-sb\") pod \"26f3f384-6499-4824-8c10-de6382e27d8b\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.367329 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-swift-storage-0\") pod \"26f3f384-6499-4824-8c10-de6382e27d8b\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.367528 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-nb\") pod \"26f3f384-6499-4824-8c10-de6382e27d8b\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.367613 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz44b\" (UniqueName: \"kubernetes.io/projected/26f3f384-6499-4824-8c10-de6382e27d8b-kube-api-access-xz44b\") pod \"26f3f384-6499-4824-8c10-de6382e27d8b\" (UID: \"26f3f384-6499-4824-8c10-de6382e27d8b\") " Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.411735 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f3f384-6499-4824-8c10-de6382e27d8b-kube-api-access-xz44b" (OuterVolumeSpecName: "kube-api-access-xz44b") pod "26f3f384-6499-4824-8c10-de6382e27d8b" (UID: "26f3f384-6499-4824-8c10-de6382e27d8b"). InnerVolumeSpecName "kube-api-access-xz44b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.473554 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz44b\" (UniqueName: \"kubernetes.io/projected/26f3f384-6499-4824-8c10-de6382e27d8b-kube-api-access-xz44b\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.525760 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "26f3f384-6499-4824-8c10-de6382e27d8b" (UID: "26f3f384-6499-4824-8c10-de6382e27d8b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.542261 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26f3f384-6499-4824-8c10-de6382e27d8b" (UID: "26f3f384-6499-4824-8c10-de6382e27d8b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.543245 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "26f3f384-6499-4824-8c10-de6382e27d8b" (UID: "26f3f384-6499-4824-8c10-de6382e27d8b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.554131 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-config" (OuterVolumeSpecName: "config") pod "26f3f384-6499-4824-8c10-de6382e27d8b" (UID: "26f3f384-6499-4824-8c10-de6382e27d8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.585365 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "26f3f384-6499-4824-8c10-de6382e27d8b" (UID: "26f3f384-6499-4824-8c10-de6382e27d8b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.596874 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.597222 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.597314 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.597407 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:16 crc kubenswrapper[4912]: I1203 00:50:16.597516 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26f3f384-6499-4824-8c10-de6382e27d8b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.033281 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" event={"ID":"04d938cc-1945-46d6-bd97-4c78397f713a","Type":"ContainerStarted","Data":"8a15618f909c98d41e2ad3cf471671a84f78374ad770cd5ad64682612cc7fa16"} Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.034152 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.057140 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" event={"ID":"26f3f384-6499-4824-8c10-de6382e27d8b","Type":"ContainerDied","Data":"7cc385ce6ba75ac85fbbee14f194beb9d3255cd4125a2b8d83bd5473791a2a5d"} Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.057214 4912 scope.go:117] "RemoveContainer" containerID="814304fa1d3f48577aeb5a367cdc99bdd58f1b535c6730312645d059b030a7ec" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.057401 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.071619 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b486667b8-shqw9" event={"ID":"10282fe1-8cff-49d3-8d01-78c23a691eac","Type":"ContainerStarted","Data":"ab9b8565fc682f2b522d55f2563bb6e7517e1dedf0d89911f612f9f985210dde"} Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.071668 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:17 crc kubenswrapper[4912]: E1203 00:50:17.074689 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" podUID="aeac5989-b199-4fcf-8d5e-c49ac6062bfd" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.172930 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.193522 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-zmrhq"] Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.215938 4912 scope.go:117] "RemoveContainer" containerID="3a505aa2448fb487f9c9d0839f071622b20dd2151dcb926fb2dd7ccf53da3a9e" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.217339 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-zmrhq"] Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.235878 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data-custom\") pod \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.236191 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data\") pod \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.236493 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-combined-ca-bundle\") pod \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.236704 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9wm9\" (UniqueName: \"kubernetes.io/projected/ef6b7b00-8f47-4d05-92f0-b43367f92c12-kube-api-access-l9wm9\") pod \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\" (UID: \"ef6b7b00-8f47-4d05-92f0-b43367f92c12\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.258760 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ef6b7b00-8f47-4d05-92f0-b43367f92c12" (UID: "ef6b7b00-8f47-4d05-92f0-b43367f92c12"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.258832 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef6b7b00-8f47-4d05-92f0-b43367f92c12-kube-api-access-l9wm9" (OuterVolumeSpecName: "kube-api-access-l9wm9") pod "ef6b7b00-8f47-4d05-92f0-b43367f92c12" (UID: "ef6b7b00-8f47-4d05-92f0-b43367f92c12"). InnerVolumeSpecName "kube-api-access-l9wm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.322544 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef6b7b00-8f47-4d05-92f0-b43367f92c12" (UID: "ef6b7b00-8f47-4d05-92f0-b43367f92c12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.332655 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data" (OuterVolumeSpecName: "config-data") pod "ef6b7b00-8f47-4d05-92f0-b43367f92c12" (UID: "ef6b7b00-8f47-4d05-92f0-b43367f92c12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.339620 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.339662 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.339672 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef6b7b00-8f47-4d05-92f0-b43367f92c12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.339682 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9wm9\" (UniqueName: \"kubernetes.io/projected/ef6b7b00-8f47-4d05-92f0-b43367f92c12-kube-api-access-l9wm9\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.416068 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.425433 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.518747 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7b79dbdd58-qznp2"] Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.538083 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69f9575dc5-s25p4"] Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.549796 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-log-httpd\") pod \"47d51262-7f59-4001-987b-f17c3b9a13dc\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.549914 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-combined-ca-bundle\") pod \"47d51262-7f59-4001-987b-f17c3b9a13dc\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550029 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-combined-ca-bundle\") pod \"5712fffb-e277-4ebd-a130-fee6f21273dc\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550155 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbdfj\" (UniqueName: \"kubernetes.io/projected/47d51262-7f59-4001-987b-f17c3b9a13dc-kube-api-access-sbdfj\") pod \"47d51262-7f59-4001-987b-f17c3b9a13dc\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550257 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data-custom\") pod \"5712fffb-e277-4ebd-a130-fee6f21273dc\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550304 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-sg-core-conf-yaml\") pod \"47d51262-7f59-4001-987b-f17c3b9a13dc\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550369 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data\") pod \"5712fffb-e277-4ebd-a130-fee6f21273dc\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550417 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-run-httpd\") pod \"47d51262-7f59-4001-987b-f17c3b9a13dc\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550512 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-scripts\") pod \"47d51262-7f59-4001-987b-f17c3b9a13dc\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550557 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnnhb\" (UniqueName: \"kubernetes.io/projected/5712fffb-e277-4ebd-a130-fee6f21273dc-kube-api-access-hnnhb\") pod \"5712fffb-e277-4ebd-a130-fee6f21273dc\" (UID: \"5712fffb-e277-4ebd-a130-fee6f21273dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550617 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-config-data\") pod \"47d51262-7f59-4001-987b-f17c3b9a13dc\" (UID: \"47d51262-7f59-4001-987b-f17c3b9a13dc\") " Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.550707 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "47d51262-7f59-4001-987b-f17c3b9a13dc" (UID: "47d51262-7f59-4001-987b-f17c3b9a13dc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.551878 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "47d51262-7f59-4001-987b-f17c3b9a13dc" (UID: "47d51262-7f59-4001-987b-f17c3b9a13dc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.553182 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.553293 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47d51262-7f59-4001-987b-f17c3b9a13dc-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.564657 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5712fffb-e277-4ebd-a130-fee6f21273dc-kube-api-access-hnnhb" (OuterVolumeSpecName: "kube-api-access-hnnhb") pod "5712fffb-e277-4ebd-a130-fee6f21273dc" (UID: "5712fffb-e277-4ebd-a130-fee6f21273dc"). InnerVolumeSpecName "kube-api-access-hnnhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.569743 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d51262-7f59-4001-987b-f17c3b9a13dc-kube-api-access-sbdfj" (OuterVolumeSpecName: "kube-api-access-sbdfj") pod "47d51262-7f59-4001-987b-f17c3b9a13dc" (UID: "47d51262-7f59-4001-987b-f17c3b9a13dc"). InnerVolumeSpecName "kube-api-access-sbdfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.569897 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-scripts" (OuterVolumeSpecName: "scripts") pod "47d51262-7f59-4001-987b-f17c3b9a13dc" (UID: "47d51262-7f59-4001-987b-f17c3b9a13dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.574418 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5712fffb-e277-4ebd-a130-fee6f21273dc" (UID: "5712fffb-e277-4ebd-a130-fee6f21273dc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.655470 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbdfj\" (UniqueName: \"kubernetes.io/projected/47d51262-7f59-4001-987b-f17c3b9a13dc-kube-api-access-sbdfj\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.659516 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.660221 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.660313 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnnhb\" (UniqueName: \"kubernetes.io/projected/5712fffb-e277-4ebd-a130-fee6f21273dc-kube-api-access-hnnhb\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.739929 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5712fffb-e277-4ebd-a130-fee6f21273dc" (UID: "5712fffb-e277-4ebd-a130-fee6f21273dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.740195 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "47d51262-7f59-4001-987b-f17c3b9a13dc" (UID: "47d51262-7f59-4001-987b-f17c3b9a13dc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.764654 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.764729 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.775137 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data" (OuterVolumeSpecName: "config-data") pod "5712fffb-e277-4ebd-a130-fee6f21273dc" (UID: "5712fffb-e277-4ebd-a130-fee6f21273dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.811636 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47d51262-7f59-4001-987b-f17c3b9a13dc" (UID: "47d51262-7f59-4001-987b-f17c3b9a13dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.859642 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-config-data" (OuterVolumeSpecName: "config-data") pod "47d51262-7f59-4001-987b-f17c3b9a13dc" (UID: "47d51262-7f59-4001-987b-f17c3b9a13dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.870427 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5712fffb-e277-4ebd-a130-fee6f21273dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.870515 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:17 crc kubenswrapper[4912]: I1203 00:50:17.870531 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d51262-7f59-4001-987b-f17c3b9a13dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.085065 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47d51262-7f59-4001-987b-f17c3b9a13dc","Type":"ContainerDied","Data":"55af0a981407746afd06137f48b16e76f720bdabd508bb06288627038a7271b0"} Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.085133 4912 scope.go:117] "RemoveContainer" containerID="c4f4514f1bd3168e24977a9dd97bf97b44cb2d4206b72a3630d9bce3c221698a" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.085281 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.090499 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b1cf8990-bf85-4de8-8121-73c43a7eccd9","Type":"ContainerStarted","Data":"27fd3d5d18405ec787f1294a45879b3235dfac7410a7de4d66d53ebbad1b6173"} Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.093865 4912 generic.go:334] "Generic (PLEG): container finished" podID="04d938cc-1945-46d6-bd97-4c78397f713a" containerID="8a15618f909c98d41e2ad3cf471671a84f78374ad770cd5ad64682612cc7fa16" exitCode=1 Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.093922 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" event={"ID":"04d938cc-1945-46d6-bd97-4c78397f713a","Type":"ContainerDied","Data":"8a15618f909c98d41e2ad3cf471671a84f78374ad770cd5ad64682612cc7fa16"} Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.097648 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-55bc76fdd7-jqgvd" event={"ID":"5712fffb-e277-4ebd-a130-fee6f21273dc","Type":"ContainerDied","Data":"d9eb7767fbf79623eee66ce885f8544cd1398163a2aa9b424a131048328b9e2f"} Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.097839 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-55bc76fdd7-jqgvd" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.098775 4912 scope.go:117] "RemoveContainer" containerID="8a15618f909c98d41e2ad3cf471671a84f78374ad770cd5ad64682612cc7fa16" Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.099486 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-694c75fc98-gbk8k_openstack(04d938cc-1945-46d6-bd97-4c78397f713a)\"" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.101869 4912 generic.go:334] "Generic (PLEG): container finished" podID="10282fe1-8cff-49d3-8d01-78c23a691eac" containerID="ab9b8565fc682f2b522d55f2563bb6e7517e1dedf0d89911f612f9f985210dde" exitCode=1 Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.102055 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b486667b8-shqw9" event={"ID":"10282fe1-8cff-49d3-8d01-78c23a691eac","Type":"ContainerDied","Data":"ab9b8565fc682f2b522d55f2563bb6e7517e1dedf0d89911f612f9f985210dde"} Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.102912 4912 scope.go:117] "RemoveContainer" containerID="ab9b8565fc682f2b522d55f2563bb6e7517e1dedf0d89911f612f9f985210dde" Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.103201 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-6b486667b8-shqw9_openstack(10282fe1-8cff-49d3-8d01-78c23a691eac)\"" pod="openstack/heat-api-6b486667b8-shqw9" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.113020 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" event={"ID":"d628e1f8-fd25-4bba-a85b-58f71eb64f58","Type":"ContainerStarted","Data":"1637ec03534bc02d4f4e665d6c348ab4bb4629ddba0c3722340a01da83c4e6d2"} Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.123970 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"64841d14-78b5-4737-90ec-b812f2994397","Type":"ContainerStarted","Data":"ab2ed2b158a1801a2127e07d42df0bc804bc45dd66823e1dac7e64f3bcf7108c"} Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.124639 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.127091 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" event={"ID":"ef6b7b00-8f47-4d05-92f0-b43367f92c12","Type":"ContainerDied","Data":"92a1770782fbf4f9c7c17e58866f4af5d740142f8296670e7c117c4d9c004a41"} Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.127157 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.128078 4912 scope.go:117] "RemoveContainer" containerID="83fcf4cdd73539bf838749b3754c62aa67a8e5cf7dcea3e10baebdb6bb591d9f" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.151738 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.710625209 podStartE2EDuration="31.151707329s" podCreationTimestamp="2025-12-03 00:49:47 +0000 UTC" firstStartedPulling="2025-12-03 00:49:48.156656068 +0000 UTC m=+1573.798676628" lastFinishedPulling="2025-12-03 00:50:16.597738178 +0000 UTC m=+1602.239758748" observedRunningTime="2025-12-03 00:50:18.148060801 +0000 UTC m=+1603.790081371" watchObservedRunningTime="2025-12-03 00:50:18.151707329 +0000 UTC m=+1603.793727889" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.152780 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69f9575dc5-s25p4" event={"ID":"c7638b15-f910-4a20-b770-c65ebdc8ac49","Type":"ContainerStarted","Data":"c51d17de32b2f70c8fe305b5bad26b757164fd30cf9d372d7bfc24ae36731b81"} Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.153115 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.164298 4912 scope.go:117] "RemoveContainer" containerID="87c8beda356bc1e9b4bfa123fc88bec6bcd22bdff115f65e76e91d068738533c" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.188828 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.213952 4912 scope.go:117] "RemoveContainer" containerID="c51247a128514615bf0495e1e754f19ebe2d6a62c0154b6b806fc34eb3ff281c" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.219516 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.240959 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.241501 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="ceilometer-central-agent" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241520 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="ceilometer-central-agent" Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.241539 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5712fffb-e277-4ebd-a130-fee6f21273dc" containerName="heat-api" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241545 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5712fffb-e277-4ebd-a130-fee6f21273dc" containerName="heat-api" Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.241559 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" containerName="dnsmasq-dns" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241565 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" containerName="dnsmasq-dns" Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.241579 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="proxy-httpd" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241585 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="proxy-httpd" Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.241599 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" containerName="heat-cfnapi" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241604 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" containerName="heat-cfnapi" Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.241615 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="ceilometer-notification-agent" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241621 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="ceilometer-notification-agent" Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.241634 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" containerName="init" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241639 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" containerName="init" Dec 03 00:50:18 crc kubenswrapper[4912]: E1203 00:50:18.241665 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="sg-core" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241673 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="sg-core" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241904 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="ceilometer-notification-agent" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241916 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="ceilometer-central-agent" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241932 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" containerName="dnsmasq-dns" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241942 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="proxy-httpd" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241951 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="5712fffb-e277-4ebd-a130-fee6f21273dc" containerName="heat-api" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241962 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" containerName="sg-core" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.241972 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" containerName="heat-cfnapi" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.245876 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.248410 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.248974 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.269064 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6d4656b7b7-t62gs"] Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.290656 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-6d4656b7b7-t62gs"] Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.296904 4912 scope.go:117] "RemoveContainer" containerID="10d7e0516e3487c09d7926fdae77e88822b063fe9b7bc934a55a48cd5cae63e0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.345555 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.390387 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.390480 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-run-httpd\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.390623 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-log-httpd\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.390847 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-scripts\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.390909 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.391247 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-config-data\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.391464 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n7bx\" (UniqueName: \"kubernetes.io/projected/2c423153-e5d1-49ca-bba7-70f2f1174226-kube-api-access-4n7bx\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.395875 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-55bc76fdd7-jqgvd"] Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.397625 4912 scope.go:117] "RemoveContainer" containerID="80efefa6f5b5386935076ff537b2269625ea692dc72a2686a831a0a282f48a08" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.409207 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-55bc76fdd7-jqgvd"] Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.416200 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=18.416166752 podStartE2EDuration="18.416166752s" podCreationTimestamp="2025-12-03 00:50:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:50:18.255169002 +0000 UTC m=+1603.897189582" watchObservedRunningTime="2025-12-03 00:50:18.416166752 +0000 UTC m=+1604.058187312" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.496118 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-69f9575dc5-s25p4" podStartSLOduration=16.496088096 podStartE2EDuration="16.496088096s" podCreationTimestamp="2025-12-03 00:50:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:50:18.350336194 +0000 UTC m=+1603.992356754" watchObservedRunningTime="2025-12-03 00:50:18.496088096 +0000 UTC m=+1604.138108656" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.496863 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-scripts\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.496931 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.496986 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-config-data\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.497014 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n7bx\" (UniqueName: \"kubernetes.io/projected/2c423153-e5d1-49ca-bba7-70f2f1174226-kube-api-access-4n7bx\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.497102 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.497126 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-run-httpd\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.497149 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-log-httpd\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.498099 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-log-httpd\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.506491 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.506662 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.506830 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-scripts\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.506941 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-run-httpd\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.509436 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-config-data\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.511628 4912 scope.go:117] "RemoveContainer" containerID="dd782341efa3233d4466a7230a183085b4b829ef804285c581314db749cd4b75" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.545224 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n7bx\" (UniqueName: \"kubernetes.io/projected/2c423153-e5d1-49ca-bba7-70f2f1174226-kube-api-access-4n7bx\") pod \"ceilometer-0\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.586137 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.586331 4912 scope.go:117] "RemoveContainer" containerID="7cbac6e5060818ac2503d5b330e04bb6e951432dda9df14f01b03998cf9aabfe" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.593781 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" path="/var/lib/kubelet/pods/26f3f384-6499-4824-8c10-de6382e27d8b/volumes" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.594886 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47d51262-7f59-4001-987b-f17c3b9a13dc" path="/var/lib/kubelet/pods/47d51262-7f59-4001-987b-f17c3b9a13dc/volumes" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.596187 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5712fffb-e277-4ebd-a130-fee6f21273dc" path="/var/lib/kubelet/pods/5712fffb-e277-4ebd-a130-fee6f21273dc/volumes" Dec 03 00:50:18 crc kubenswrapper[4912]: I1203 00:50:18.596798 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" path="/var/lib/kubelet/pods/ef6b7b00-8f47-4d05-92f0-b43367f92c12/volumes" Dec 03 00:50:19 crc kubenswrapper[4912]: I1203 00:50:19.106475 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:19 crc kubenswrapper[4912]: W1203 00:50:19.122775 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c423153_e5d1_49ca_bba7_70f2f1174226.slice/crio-d7725108d59ee5d6f220f41a00469f23e36ec0e63448e9d95138aed6b420d1c6 WatchSource:0}: Error finding container d7725108d59ee5d6f220f41a00469f23e36ec0e63448e9d95138aed6b420d1c6: Status 404 returned error can't find the container with id d7725108d59ee5d6f220f41a00469f23e36ec0e63448e9d95138aed6b420d1c6 Dec 03 00:50:19 crc kubenswrapper[4912]: I1203 00:50:19.169325 4912 scope.go:117] "RemoveContainer" containerID="ab9b8565fc682f2b522d55f2563bb6e7517e1dedf0d89911f612f9f985210dde" Dec 03 00:50:19 crc kubenswrapper[4912]: E1203 00:50:19.169638 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-6b486667b8-shqw9_openstack(10282fe1-8cff-49d3-8d01-78c23a691eac)\"" pod="openstack/heat-api-6b486667b8-shqw9" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" Dec 03 00:50:19 crc kubenswrapper[4912]: I1203 00:50:19.170737 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" event={"ID":"d628e1f8-fd25-4bba-a85b-58f71eb64f58","Type":"ContainerStarted","Data":"bfd01aa91c6d8b8d3eec8616ba5eccf009db870e962ac170efa084202daab34d"} Dec 03 00:50:19 crc kubenswrapper[4912]: I1203 00:50:19.172051 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:19 crc kubenswrapper[4912]: I1203 00:50:19.178370 4912 scope.go:117] "RemoveContainer" containerID="8a15618f909c98d41e2ad3cf471671a84f78374ad770cd5ad64682612cc7fa16" Dec 03 00:50:19 crc kubenswrapper[4912]: E1203 00:50:19.178726 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-694c75fc98-gbk8k_openstack(04d938cc-1945-46d6-bd97-4c78397f713a)\"" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" Dec 03 00:50:19 crc kubenswrapper[4912]: I1203 00:50:19.184482 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerStarted","Data":"d7725108d59ee5d6f220f41a00469f23e36ec0e63448e9d95138aed6b420d1c6"} Dec 03 00:50:19 crc kubenswrapper[4912]: I1203 00:50:19.189521 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69f9575dc5-s25p4" event={"ID":"c7638b15-f910-4a20-b770-c65ebdc8ac49","Type":"ContainerStarted","Data":"c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488"} Dec 03 00:50:19 crc kubenswrapper[4912]: I1203 00:50:19.245244 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" podStartSLOduration=16.245216953 podStartE2EDuration="16.245216953s" podCreationTimestamp="2025-12-03 00:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:50:19.244330699 +0000 UTC m=+1604.886351269" watchObservedRunningTime="2025-12-03 00:50:19.245216953 +0000 UTC m=+1604.887237513" Dec 03 00:50:20 crc kubenswrapper[4912]: I1203 00:50:20.183615 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:20 crc kubenswrapper[4912]: I1203 00:50:20.203770 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:20 crc kubenswrapper[4912]: I1203 00:50:20.212720 4912 scope.go:117] "RemoveContainer" containerID="ab9b8565fc682f2b522d55f2563bb6e7517e1dedf0d89911f612f9f985210dde" Dec 03 00:50:20 crc kubenswrapper[4912]: E1203 00:50:20.213037 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-6b486667b8-shqw9_openstack(10282fe1-8cff-49d3-8d01-78c23a691eac)\"" pod="openstack/heat-api-6b486667b8-shqw9" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" Dec 03 00:50:20 crc kubenswrapper[4912]: I1203 00:50:20.213103 4912 scope.go:117] "RemoveContainer" containerID="8a15618f909c98d41e2ad3cf471671a84f78374ad770cd5ad64682612cc7fa16" Dec 03 00:50:20 crc kubenswrapper[4912]: E1203 00:50:20.213366 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-694c75fc98-gbk8k_openstack(04d938cc-1945-46d6-bd97-4c78397f713a)\"" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" Dec 03 00:50:20 crc kubenswrapper[4912]: I1203 00:50:20.232870 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:50:20 crc kubenswrapper[4912]: I1203 00:50:20.302489 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-78c76df587-7nwws"] Dec 03 00:50:20 crc kubenswrapper[4912]: I1203 00:50:20.303296 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-78c76df587-7nwws" podUID="4881ee51-d772-4954-8e62-29584cc7b3c8" containerName="heat-engine" containerID="cri-o://2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5" gracePeriod=60 Dec 03 00:50:20 crc kubenswrapper[4912]: I1203 00:50:20.557772 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6578955fd5-zmrhq" podUID="26f3f384-6499-4824-8c10-de6382e27d8b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.185:5353: i/o timeout" Dec 03 00:50:21 crc kubenswrapper[4912]: I1203 00:50:21.225324 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerStarted","Data":"c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f"} Dec 03 00:50:21 crc kubenswrapper[4912]: I1203 00:50:21.225906 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerStarted","Data":"6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed"} Dec 03 00:50:21 crc kubenswrapper[4912]: E1203 00:50:21.344926 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 00:50:21 crc kubenswrapper[4912]: E1203 00:50:21.346320 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 00:50:21 crc kubenswrapper[4912]: E1203 00:50:21.347990 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 00:50:21 crc kubenswrapper[4912]: E1203 00:50:21.348091 4912 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-78c76df587-7nwws" podUID="4881ee51-d772-4954-8e62-29584cc7b3c8" containerName="heat-engine" Dec 03 00:50:21 crc kubenswrapper[4912]: I1203 00:50:21.851579 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:21 crc kubenswrapper[4912]: I1203 00:50:21.999877 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rz5pb"] Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.003759 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.046551 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz5pb"] Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.093485 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-catalog-content\") pod \"redhat-marketplace-rz5pb\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.093575 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-utilities\") pod \"redhat-marketplace-rz5pb\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.093652 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sx75\" (UniqueName: \"kubernetes.io/projected/9f75e609-070b-43bc-b217-c1f503cb9360-kube-api-access-6sx75\") pod \"redhat-marketplace-rz5pb\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.195545 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-utilities\") pod \"redhat-marketplace-rz5pb\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.195686 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sx75\" (UniqueName: \"kubernetes.io/projected/9f75e609-070b-43bc-b217-c1f503cb9360-kube-api-access-6sx75\") pod \"redhat-marketplace-rz5pb\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.195805 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-catalog-content\") pod \"redhat-marketplace-rz5pb\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.196017 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-utilities\") pod \"redhat-marketplace-rz5pb\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.196336 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-catalog-content\") pod \"redhat-marketplace-rz5pb\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.217729 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sx75\" (UniqueName: \"kubernetes.io/projected/9f75e609-070b-43bc-b217-c1f503cb9360-kube-api-access-6sx75\") pod \"redhat-marketplace-rz5pb\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.383281 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:22 crc kubenswrapper[4912]: I1203 00:50:22.976891 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz5pb"] Dec 03 00:50:23 crc kubenswrapper[4912]: I1203 00:50:23.250337 4912 generic.go:334] "Generic (PLEG): container finished" podID="9f75e609-070b-43bc-b217-c1f503cb9360" containerID="0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a" exitCode=0 Dec 03 00:50:23 crc kubenswrapper[4912]: I1203 00:50:23.250527 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz5pb" event={"ID":"9f75e609-070b-43bc-b217-c1f503cb9360","Type":"ContainerDied","Data":"0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a"} Dec 03 00:50:23 crc kubenswrapper[4912]: I1203 00:50:23.250965 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz5pb" event={"ID":"9f75e609-070b-43bc-b217-c1f503cb9360","Type":"ContainerStarted","Data":"c59210b1d712ec0b5f56d6089acc8975784b9eac977bf1ef163c34eaeaa220a9"} Dec 03 00:50:23 crc kubenswrapper[4912]: I1203 00:50:23.253537 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerStarted","Data":"42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b"} Dec 03 00:50:24 crc kubenswrapper[4912]: I1203 00:50:24.270595 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz5pb" event={"ID":"9f75e609-070b-43bc-b217-c1f503cb9360","Type":"ContainerStarted","Data":"51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5"} Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.218811 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.295030 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6b486667b8-shqw9"] Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.346573 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerStarted","Data":"1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf"} Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.347044 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="ceilometer-central-agent" containerID="cri-o://6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed" gracePeriod=30 Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.347071 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.347223 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="proxy-httpd" containerID="cri-o://1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf" gracePeriod=30 Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.347271 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="sg-core" containerID="cri-o://42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b" gracePeriod=30 Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.347300 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="ceilometer-notification-agent" containerID="cri-o://c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f" gracePeriod=30 Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.377731 4912 generic.go:334] "Generic (PLEG): container finished" podID="9f75e609-070b-43bc-b217-c1f503cb9360" containerID="51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5" exitCode=0 Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.377808 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz5pb" event={"ID":"9f75e609-070b-43bc-b217-c1f503cb9360","Type":"ContainerDied","Data":"51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5"} Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.414272 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.761404669 podStartE2EDuration="7.414239135s" podCreationTimestamp="2025-12-03 00:50:18 +0000 UTC" firstStartedPulling="2025-12-03 00:50:19.126141533 +0000 UTC m=+1604.768162103" lastFinishedPulling="2025-12-03 00:50:24.778976019 +0000 UTC m=+1610.420996569" observedRunningTime="2025-12-03 00:50:25.384345976 +0000 UTC m=+1611.026366546" watchObservedRunningTime="2025-12-03 00:50:25.414239135 +0000 UTC m=+1611.056259695" Dec 03 00:50:25 crc kubenswrapper[4912]: I1203 00:50:25.932840 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.045485 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtfw4\" (UniqueName: \"kubernetes.io/projected/10282fe1-8cff-49d3-8d01-78c23a691eac-kube-api-access-xtfw4\") pod \"10282fe1-8cff-49d3-8d01-78c23a691eac\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.045931 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-combined-ca-bundle\") pod \"10282fe1-8cff-49d3-8d01-78c23a691eac\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.046064 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data\") pod \"10282fe1-8cff-49d3-8d01-78c23a691eac\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.046120 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data-custom\") pod \"10282fe1-8cff-49d3-8d01-78c23a691eac\" (UID: \"10282fe1-8cff-49d3-8d01-78c23a691eac\") " Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.061284 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "10282fe1-8cff-49d3-8d01-78c23a691eac" (UID: "10282fe1-8cff-49d3-8d01-78c23a691eac"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.065326 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10282fe1-8cff-49d3-8d01-78c23a691eac-kube-api-access-xtfw4" (OuterVolumeSpecName: "kube-api-access-xtfw4") pod "10282fe1-8cff-49d3-8d01-78c23a691eac" (UID: "10282fe1-8cff-49d3-8d01-78c23a691eac"). InnerVolumeSpecName "kube-api-access-xtfw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.114618 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10282fe1-8cff-49d3-8d01-78c23a691eac" (UID: "10282fe1-8cff-49d3-8d01-78c23a691eac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.153642 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtfw4\" (UniqueName: \"kubernetes.io/projected/10282fe1-8cff-49d3-8d01-78c23a691eac-kube-api-access-xtfw4\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.153693 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.153709 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.162631 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data" (OuterVolumeSpecName: "config-data") pod "10282fe1-8cff-49d3-8d01-78c23a691eac" (UID: "10282fe1-8cff-49d3-8d01-78c23a691eac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.235221 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.259679 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10282fe1-8cff-49d3-8d01-78c23a691eac-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.312750 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-694c75fc98-gbk8k"] Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.436222 4912 generic.go:334] "Generic (PLEG): container finished" podID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerID="1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf" exitCode=0 Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.436596 4912 generic.go:334] "Generic (PLEG): container finished" podID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerID="42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b" exitCode=2 Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.436604 4912 generic.go:334] "Generic (PLEG): container finished" podID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerID="c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f" exitCode=0 Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.436707 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerDied","Data":"1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf"} Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.436743 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerDied","Data":"42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b"} Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.436754 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerDied","Data":"c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f"} Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.442311 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b486667b8-shqw9" event={"ID":"10282fe1-8cff-49d3-8d01-78c23a691eac","Type":"ContainerDied","Data":"61d9ee2ae71b22b31f025a67a67a908974d3505f5ac2a1b0a5e46f5977c84a16"} Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.442376 4912 scope.go:117] "RemoveContainer" containerID="ab9b8565fc682f2b522d55f2563bb6e7517e1dedf0d89911f612f9f985210dde" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.442586 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b486667b8-shqw9" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.510496 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6b486667b8-shqw9"] Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.524307 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6b486667b8-shqw9"] Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.604816 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" path="/var/lib/kubelet/pods/10282fe1-8cff-49d3-8d01-78c23a691eac/volumes" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.739418 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.854219 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-6d4656b7b7-t62gs" podUID="ef6b7b00-8f47-4d05-92f0-b43367f92c12" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.200:8000/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.874389 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data\") pod \"04d938cc-1945-46d6-bd97-4c78397f713a\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.874478 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dlp8\" (UniqueName: \"kubernetes.io/projected/04d938cc-1945-46d6-bd97-4c78397f713a-kube-api-access-6dlp8\") pod \"04d938cc-1945-46d6-bd97-4c78397f713a\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.874748 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-combined-ca-bundle\") pod \"04d938cc-1945-46d6-bd97-4c78397f713a\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.874848 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data-custom\") pod \"04d938cc-1945-46d6-bd97-4c78397f713a\" (UID: \"04d938cc-1945-46d6-bd97-4c78397f713a\") " Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.884667 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04d938cc-1945-46d6-bd97-4c78397f713a-kube-api-access-6dlp8" (OuterVolumeSpecName: "kube-api-access-6dlp8") pod "04d938cc-1945-46d6-bd97-4c78397f713a" (UID: "04d938cc-1945-46d6-bd97-4c78397f713a"). InnerVolumeSpecName "kube-api-access-6dlp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.888694 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "04d938cc-1945-46d6-bd97-4c78397f713a" (UID: "04d938cc-1945-46d6-bd97-4c78397f713a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.929538 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04d938cc-1945-46d6-bd97-4c78397f713a" (UID: "04d938cc-1945-46d6-bd97-4c78397f713a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.968331 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data" (OuterVolumeSpecName: "config-data") pod "04d938cc-1945-46d6-bd97-4c78397f713a" (UID: "04d938cc-1945-46d6-bd97-4c78397f713a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.977373 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.977455 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.977467 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04d938cc-1945-46d6-bd97-4c78397f713a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.977478 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dlp8\" (UniqueName: \"kubernetes.io/projected/04d938cc-1945-46d6-bd97-4c78397f713a-kube-api-access-6dlp8\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:26 crc kubenswrapper[4912]: I1203 00:50:26.992492 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-55bc76fdd7-jqgvd" podUID="5712fffb-e277-4ebd-a130-fee6f21273dc" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.201:8004/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.431996 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.486608 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz5pb" event={"ID":"9f75e609-070b-43bc-b217-c1f503cb9360","Type":"ContainerStarted","Data":"3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d"} Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.489283 4912 generic.go:334] "Generic (PLEG): container finished" podID="4881ee51-d772-4954-8e62-29584cc7b3c8" containerID="2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5" exitCode=0 Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.489491 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-78c76df587-7nwws" event={"ID":"4881ee51-d772-4954-8e62-29584cc7b3c8","Type":"ContainerDied","Data":"2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5"} Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.489591 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-78c76df587-7nwws" event={"ID":"4881ee51-d772-4954-8e62-29584cc7b3c8","Type":"ContainerDied","Data":"75aa1ff0e847c2a0044814c742acc65f0a4f881516e5a266413e7892969462b7"} Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.489686 4912 scope.go:117] "RemoveContainer" containerID="2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.489944 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-78c76df587-7nwws" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.521228 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" event={"ID":"04d938cc-1945-46d6-bd97-4c78397f713a","Type":"ContainerDied","Data":"2343dc29e4732fade042928207d1638357ecb48b88af44db995404019f245238"} Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.521950 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-694c75fc98-gbk8k" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.528365 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rz5pb" podStartSLOduration=3.9236162930000003 podStartE2EDuration="6.528339025s" podCreationTimestamp="2025-12-03 00:50:21 +0000 UTC" firstStartedPulling="2025-12-03 00:50:23.25688622 +0000 UTC m=+1608.898906780" lastFinishedPulling="2025-12-03 00:50:25.861608952 +0000 UTC m=+1611.503629512" observedRunningTime="2025-12-03 00:50:27.522777067 +0000 UTC m=+1613.164797637" watchObservedRunningTime="2025-12-03 00:50:27.528339025 +0000 UTC m=+1613.170359585" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.553594 4912 scope.go:117] "RemoveContainer" containerID="2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5" Dec 03 00:50:27 crc kubenswrapper[4912]: E1203 00:50:27.554516 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5\": container with ID starting with 2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5 not found: ID does not exist" containerID="2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.554556 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5"} err="failed to get container status \"2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5\": rpc error: code = NotFound desc = could not find container \"2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5\": container with ID starting with 2aa4ae21366ed9386b550e3fe8da1ae2b6dad17e239176e395925a30007bbdc5 not found: ID does not exist" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.554583 4912 scope.go:117] "RemoveContainer" containerID="8a15618f909c98d41e2ad3cf471671a84f78374ad770cd5ad64682612cc7fa16" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.592276 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data-custom\") pod \"4881ee51-d772-4954-8e62-29584cc7b3c8\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.593825 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-combined-ca-bundle\") pod \"4881ee51-d772-4954-8e62-29584cc7b3c8\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.594045 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data\") pod \"4881ee51-d772-4954-8e62-29584cc7b3c8\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.594932 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr28t\" (UniqueName: \"kubernetes.io/projected/4881ee51-d772-4954-8e62-29584cc7b3c8-kube-api-access-zr28t\") pod \"4881ee51-d772-4954-8e62-29584cc7b3c8\" (UID: \"4881ee51-d772-4954-8e62-29584cc7b3c8\") " Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.598819 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-694c75fc98-gbk8k"] Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.602954 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4881ee51-d772-4954-8e62-29584cc7b3c8" (UID: "4881ee51-d772-4954-8e62-29584cc7b3c8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.609419 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-694c75fc98-gbk8k"] Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.617728 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4881ee51-d772-4954-8e62-29584cc7b3c8-kube-api-access-zr28t" (OuterVolumeSpecName: "kube-api-access-zr28t") pod "4881ee51-d772-4954-8e62-29584cc7b3c8" (UID: "4881ee51-d772-4954-8e62-29584cc7b3c8"). InnerVolumeSpecName "kube-api-access-zr28t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.652398 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4881ee51-d772-4954-8e62-29584cc7b3c8" (UID: "4881ee51-d772-4954-8e62-29584cc7b3c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.687574 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data" (OuterVolumeSpecName: "config-data") pod "4881ee51-d772-4954-8e62-29584cc7b3c8" (UID: "4881ee51-d772-4954-8e62-29584cc7b3c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.701447 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.701488 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr28t\" (UniqueName: \"kubernetes.io/projected/4881ee51-d772-4954-8e62-29584cc7b3c8-kube-api-access-zr28t\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.701498 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.701510 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4881ee51-d772-4954-8e62-29584cc7b3c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.827483 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-78c76df587-7nwws"] Dec 03 00:50:27 crc kubenswrapper[4912]: I1203 00:50:27.844650 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-78c76df587-7nwws"] Dec 03 00:50:28 crc kubenswrapper[4912]: I1203 00:50:28.544509 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" event={"ID":"aeac5989-b199-4fcf-8d5e-c49ac6062bfd","Type":"ContainerStarted","Data":"d8d0315b779492e8da1aad0ac6bd679a8db367aa217f9b35baff966ae1e97134"} Dec 03 00:50:28 crc kubenswrapper[4912]: I1203 00:50:28.566302 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" podStartSLOduration=2.685799117 podStartE2EDuration="30.566281505s" podCreationTimestamp="2025-12-03 00:49:58 +0000 UTC" firstStartedPulling="2025-12-03 00:50:00.181806737 +0000 UTC m=+1585.823827297" lastFinishedPulling="2025-12-03 00:50:28.062289125 +0000 UTC m=+1613.704309685" observedRunningTime="2025-12-03 00:50:28.564385395 +0000 UTC m=+1614.206405975" watchObservedRunningTime="2025-12-03 00:50:28.566281505 +0000 UTC m=+1614.208302065" Dec 03 00:50:28 crc kubenswrapper[4912]: I1203 00:50:28.587387 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" path="/var/lib/kubelet/pods/04d938cc-1945-46d6-bd97-4c78397f713a/volumes" Dec 03 00:50:28 crc kubenswrapper[4912]: I1203 00:50:28.588283 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4881ee51-d772-4954-8e62-29584cc7b3c8" path="/var/lib/kubelet/pods/4881ee51-d772-4954-8e62-29584cc7b3c8/volumes" Dec 03 00:50:29 crc kubenswrapper[4912]: I1203 00:50:29.059046 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 00:50:29 crc kubenswrapper[4912]: I1203 00:50:29.572106 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:50:29 crc kubenswrapper[4912]: E1203 00:50:29.573107 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.383895 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.384987 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.439480 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.471084 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-px6bb"] Dec 03 00:50:32 crc kubenswrapper[4912]: E1203 00:50:32.471870 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4881ee51-d772-4954-8e62-29584cc7b3c8" containerName="heat-engine" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.471894 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="4881ee51-d772-4954-8e62-29584cc7b3c8" containerName="heat-engine" Dec 03 00:50:32 crc kubenswrapper[4912]: E1203 00:50:32.471916 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" containerName="heat-api" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.471924 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" containerName="heat-api" Dec 03 00:50:32 crc kubenswrapper[4912]: E1203 00:50:32.471964 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" containerName="heat-cfnapi" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.471973 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" containerName="heat-cfnapi" Dec 03 00:50:32 crc kubenswrapper[4912]: E1203 00:50:32.472002 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" containerName="heat-cfnapi" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.472012 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" containerName="heat-cfnapi" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.472296 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="4881ee51-d772-4954-8e62-29584cc7b3c8" containerName="heat-engine" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.472329 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" containerName="heat-api" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.472342 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" containerName="heat-cfnapi" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.472352 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" containerName="heat-api" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.472369 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="04d938cc-1945-46d6-bd97-4c78397f713a" containerName="heat-cfnapi" Dec 03 00:50:32 crc kubenswrapper[4912]: E1203 00:50:32.472593 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" containerName="heat-api" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.472607 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="10282fe1-8cff-49d3-8d01-78c23a691eac" containerName="heat-api" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.474260 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.490613 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-px6bb"] Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.625302 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-utilities\") pod \"certified-operators-px6bb\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.625390 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdkl8\" (UniqueName: \"kubernetes.io/projected/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-kube-api-access-cdkl8\") pod \"certified-operators-px6bb\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.626021 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-catalog-content\") pod \"certified-operators-px6bb\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.672641 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.728751 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-catalog-content\") pod \"certified-operators-px6bb\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.728886 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-utilities\") pod \"certified-operators-px6bb\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.728933 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdkl8\" (UniqueName: \"kubernetes.io/projected/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-kube-api-access-cdkl8\") pod \"certified-operators-px6bb\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.730152 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-utilities\") pod \"certified-operators-px6bb\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.730246 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-catalog-content\") pod \"certified-operators-px6bb\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.752329 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdkl8\" (UniqueName: \"kubernetes.io/projected/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-kube-api-access-cdkl8\") pod \"certified-operators-px6bb\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:32 crc kubenswrapper[4912]: I1203 00:50:32.845075 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:33 crc kubenswrapper[4912]: I1203 00:50:33.413393 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-px6bb"] Dec 03 00:50:33 crc kubenswrapper[4912]: W1203 00:50:33.418638 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfab2bf1c_1cda_4c59_a54b_c24db7ec4ff9.slice/crio-4ee09abc49805ffcc5d7538bef971d3abbf4b97231ea91ecd3d379e98e118019 WatchSource:0}: Error finding container 4ee09abc49805ffcc5d7538bef971d3abbf4b97231ea91ecd3d379e98e118019: Status 404 returned error can't find the container with id 4ee09abc49805ffcc5d7538bef971d3abbf4b97231ea91ecd3d379e98e118019 Dec 03 00:50:33 crc kubenswrapper[4912]: I1203 00:50:33.630730 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-px6bb" event={"ID":"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9","Type":"ContainerStarted","Data":"4ee09abc49805ffcc5d7538bef971d3abbf4b97231ea91ecd3d379e98e118019"} Dec 03 00:50:34 crc kubenswrapper[4912]: I1203 00:50:34.651863 4912 generic.go:334] "Generic (PLEG): container finished" podID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerID="0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843" exitCode=0 Dec 03 00:50:34 crc kubenswrapper[4912]: I1203 00:50:34.652996 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-px6bb" event={"ID":"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9","Type":"ContainerDied","Data":"0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843"} Dec 03 00:50:34 crc kubenswrapper[4912]: I1203 00:50:34.830421 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz5pb"] Dec 03 00:50:34 crc kubenswrapper[4912]: I1203 00:50:34.830739 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rz5pb" podUID="9f75e609-070b-43bc-b217-c1f503cb9360" containerName="registry-server" containerID="cri-o://3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d" gracePeriod=2 Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.396535 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.505672 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-catalog-content\") pod \"9f75e609-070b-43bc-b217-c1f503cb9360\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.506027 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-utilities\") pod \"9f75e609-070b-43bc-b217-c1f503cb9360\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.506064 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sx75\" (UniqueName: \"kubernetes.io/projected/9f75e609-070b-43bc-b217-c1f503cb9360-kube-api-access-6sx75\") pod \"9f75e609-070b-43bc-b217-c1f503cb9360\" (UID: \"9f75e609-070b-43bc-b217-c1f503cb9360\") " Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.506869 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-utilities" (OuterVolumeSpecName: "utilities") pod "9f75e609-070b-43bc-b217-c1f503cb9360" (UID: "9f75e609-070b-43bc-b217-c1f503cb9360"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.507093 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.518225 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f75e609-070b-43bc-b217-c1f503cb9360-kube-api-access-6sx75" (OuterVolumeSpecName: "kube-api-access-6sx75") pod "9f75e609-070b-43bc-b217-c1f503cb9360" (UID: "9f75e609-070b-43bc-b217-c1f503cb9360"). InnerVolumeSpecName "kube-api-access-6sx75". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.528605 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f75e609-070b-43bc-b217-c1f503cb9360" (UID: "9f75e609-070b-43bc-b217-c1f503cb9360"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.610158 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sx75\" (UniqueName: \"kubernetes.io/projected/9f75e609-070b-43bc-b217-c1f503cb9360-kube-api-access-6sx75\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.610200 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f75e609-070b-43bc-b217-c1f503cb9360-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.667368 4912 generic.go:334] "Generic (PLEG): container finished" podID="9f75e609-070b-43bc-b217-c1f503cb9360" containerID="3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d" exitCode=0 Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.667905 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz5pb" event={"ID":"9f75e609-070b-43bc-b217-c1f503cb9360","Type":"ContainerDied","Data":"3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d"} Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.667951 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rz5pb" event={"ID":"9f75e609-070b-43bc-b217-c1f503cb9360","Type":"ContainerDied","Data":"c59210b1d712ec0b5f56d6089acc8975784b9eac977bf1ef163c34eaeaa220a9"} Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.667979 4912 scope.go:117] "RemoveContainer" containerID="3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.668184 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rz5pb" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.716641 4912 scope.go:117] "RemoveContainer" containerID="51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.720871 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz5pb"] Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.732275 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rz5pb"] Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.749368 4912 scope.go:117] "RemoveContainer" containerID="0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.779376 4912 scope.go:117] "RemoveContainer" containerID="3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d" Dec 03 00:50:35 crc kubenswrapper[4912]: E1203 00:50:35.779981 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d\": container with ID starting with 3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d not found: ID does not exist" containerID="3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.780029 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d"} err="failed to get container status \"3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d\": rpc error: code = NotFound desc = could not find container \"3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d\": container with ID starting with 3b7377aad84fe9b873a91e0207b556b03622465c71341c21ccec14c33a3bc79d not found: ID does not exist" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.780060 4912 scope.go:117] "RemoveContainer" containerID="51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5" Dec 03 00:50:35 crc kubenswrapper[4912]: E1203 00:50:35.780478 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5\": container with ID starting with 51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5 not found: ID does not exist" containerID="51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.780549 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5"} err="failed to get container status \"51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5\": rpc error: code = NotFound desc = could not find container \"51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5\": container with ID starting with 51dcf6581a7ca740e0b8101654487bf7f6834e2563467c818e553445fc4456e5 not found: ID does not exist" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.780604 4912 scope.go:117] "RemoveContainer" containerID="0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a" Dec 03 00:50:35 crc kubenswrapper[4912]: E1203 00:50:35.781164 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a\": container with ID starting with 0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a not found: ID does not exist" containerID="0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a" Dec 03 00:50:35 crc kubenswrapper[4912]: I1203 00:50:35.781214 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a"} err="failed to get container status \"0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a\": rpc error: code = NotFound desc = could not find container \"0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a\": container with ID starting with 0aa58ee5b44d3e476c0061fe7cfa29ea3ad90d45c796848f42feaf07e165722a not found: ID does not exist" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.585997 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f75e609-070b-43bc-b217-c1f503cb9360" path="/var/lib/kubelet/pods/9f75e609-070b-43bc-b217-c1f503cb9360/volumes" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.603371 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.684374 4912 generic.go:334] "Generic (PLEG): container finished" podID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerID="5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17" exitCode=0 Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.684524 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-px6bb" event={"ID":"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9","Type":"ContainerDied","Data":"5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17"} Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.689227 4912 generic.go:334] "Generic (PLEG): container finished" podID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerID="6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed" exitCode=0 Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.689300 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerDied","Data":"6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed"} Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.689337 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2c423153-e5d1-49ca-bba7-70f2f1174226","Type":"ContainerDied","Data":"d7725108d59ee5d6f220f41a00469f23e36ec0e63448e9d95138aed6b420d1c6"} Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.689356 4912 scope.go:117] "RemoveContainer" containerID="1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.689517 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.740884 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-combined-ca-bundle\") pod \"2c423153-e5d1-49ca-bba7-70f2f1174226\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.740974 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n7bx\" (UniqueName: \"kubernetes.io/projected/2c423153-e5d1-49ca-bba7-70f2f1174226-kube-api-access-4n7bx\") pod \"2c423153-e5d1-49ca-bba7-70f2f1174226\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.741017 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-log-httpd\") pod \"2c423153-e5d1-49ca-bba7-70f2f1174226\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.741094 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-scripts\") pod \"2c423153-e5d1-49ca-bba7-70f2f1174226\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.741139 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-run-httpd\") pod \"2c423153-e5d1-49ca-bba7-70f2f1174226\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.741194 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-config-data\") pod \"2c423153-e5d1-49ca-bba7-70f2f1174226\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.741462 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-sg-core-conf-yaml\") pod \"2c423153-e5d1-49ca-bba7-70f2f1174226\" (UID: \"2c423153-e5d1-49ca-bba7-70f2f1174226\") " Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.743143 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2c423153-e5d1-49ca-bba7-70f2f1174226" (UID: "2c423153-e5d1-49ca-bba7-70f2f1174226"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.744614 4912 scope.go:117] "RemoveContainer" containerID="42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.746445 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2c423153-e5d1-49ca-bba7-70f2f1174226" (UID: "2c423153-e5d1-49ca-bba7-70f2f1174226"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.765464 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-scripts" (OuterVolumeSpecName: "scripts") pod "2c423153-e5d1-49ca-bba7-70f2f1174226" (UID: "2c423153-e5d1-49ca-bba7-70f2f1174226"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.765927 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c423153-e5d1-49ca-bba7-70f2f1174226-kube-api-access-4n7bx" (OuterVolumeSpecName: "kube-api-access-4n7bx") pod "2c423153-e5d1-49ca-bba7-70f2f1174226" (UID: "2c423153-e5d1-49ca-bba7-70f2f1174226"). InnerVolumeSpecName "kube-api-access-4n7bx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.783062 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2c423153-e5d1-49ca-bba7-70f2f1174226" (UID: "2c423153-e5d1-49ca-bba7-70f2f1174226"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.836090 4912 scope.go:117] "RemoveContainer" containerID="c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.845633 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n7bx\" (UniqueName: \"kubernetes.io/projected/2c423153-e5d1-49ca-bba7-70f2f1174226-kube-api-access-4n7bx\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.845663 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.845676 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.845687 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2c423153-e5d1-49ca-bba7-70f2f1174226-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.845696 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.850073 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c423153-e5d1-49ca-bba7-70f2f1174226" (UID: "2c423153-e5d1-49ca-bba7-70f2f1174226"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.865449 4912 scope.go:117] "RemoveContainer" containerID="6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.879833 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-config-data" (OuterVolumeSpecName: "config-data") pod "2c423153-e5d1-49ca-bba7-70f2f1174226" (UID: "2c423153-e5d1-49ca-bba7-70f2f1174226"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.889249 4912 scope.go:117] "RemoveContainer" containerID="1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf" Dec 03 00:50:36 crc kubenswrapper[4912]: E1203 00:50:36.890114 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf\": container with ID starting with 1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf not found: ID does not exist" containerID="1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.890184 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf"} err="failed to get container status \"1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf\": rpc error: code = NotFound desc = could not find container \"1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf\": container with ID starting with 1eff81e19fbc2c56bc2ca020546153492ed3216d927b2cb3fac50a7fa153f9bf not found: ID does not exist" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.890226 4912 scope.go:117] "RemoveContainer" containerID="42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b" Dec 03 00:50:36 crc kubenswrapper[4912]: E1203 00:50:36.892361 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b\": container with ID starting with 42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b not found: ID does not exist" containerID="42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.892412 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b"} err="failed to get container status \"42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b\": rpc error: code = NotFound desc = could not find container \"42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b\": container with ID starting with 42a222d82a69b45959ec5179b9089466d7da12a36b37f7f3b84289a8ce2ae02b not found: ID does not exist" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.892458 4912 scope.go:117] "RemoveContainer" containerID="c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f" Dec 03 00:50:36 crc kubenswrapper[4912]: E1203 00:50:36.893122 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f\": container with ID starting with c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f not found: ID does not exist" containerID="c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.893149 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f"} err="failed to get container status \"c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f\": rpc error: code = NotFound desc = could not find container \"c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f\": container with ID starting with c8c3df51c39fe4387d40100bddf43b1f8019c7f313d097241f1897b5b26c0f1f not found: ID does not exist" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.893166 4912 scope.go:117] "RemoveContainer" containerID="6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed" Dec 03 00:50:36 crc kubenswrapper[4912]: E1203 00:50:36.893466 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed\": container with ID starting with 6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed not found: ID does not exist" containerID="6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.893486 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed"} err="failed to get container status \"6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed\": rpc error: code = NotFound desc = could not find container \"6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed\": container with ID starting with 6cdfa6cc315173b6aefc20de74303beb328a32dd40e77ab64b772b8d29ce90ed not found: ID does not exist" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.948603 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:36 crc kubenswrapper[4912]: I1203 00:50:36.948649 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c423153-e5d1-49ca-bba7-70f2f1174226-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.028920 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.041849 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.075021 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:37 crc kubenswrapper[4912]: E1203 00:50:37.075842 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="ceilometer-notification-agent" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.075876 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="ceilometer-notification-agent" Dec 03 00:50:37 crc kubenswrapper[4912]: E1203 00:50:37.075891 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="sg-core" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.075901 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="sg-core" Dec 03 00:50:37 crc kubenswrapper[4912]: E1203 00:50:37.075950 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f75e609-070b-43bc-b217-c1f503cb9360" containerName="extract-utilities" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.075959 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f75e609-070b-43bc-b217-c1f503cb9360" containerName="extract-utilities" Dec 03 00:50:37 crc kubenswrapper[4912]: E1203 00:50:37.075979 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="proxy-httpd" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.075984 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="proxy-httpd" Dec 03 00:50:37 crc kubenswrapper[4912]: E1203 00:50:37.075997 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="ceilometer-central-agent" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.076003 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="ceilometer-central-agent" Dec 03 00:50:37 crc kubenswrapper[4912]: E1203 00:50:37.076020 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f75e609-070b-43bc-b217-c1f503cb9360" containerName="registry-server" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.076026 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f75e609-070b-43bc-b217-c1f503cb9360" containerName="registry-server" Dec 03 00:50:37 crc kubenswrapper[4912]: E1203 00:50:37.076035 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f75e609-070b-43bc-b217-c1f503cb9360" containerName="extract-content" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.076041 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f75e609-070b-43bc-b217-c1f503cb9360" containerName="extract-content" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.076265 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f75e609-070b-43bc-b217-c1f503cb9360" containerName="registry-server" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.076277 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="ceilometer-central-agent" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.076290 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="proxy-httpd" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.076303 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="ceilometer-notification-agent" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.076310 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" containerName="sg-core" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.078771 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.081310 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.081614 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.092267 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.154453 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-log-httpd\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.154536 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.154575 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-scripts\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.154614 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-run-httpd\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.154649 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.154688 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jpvb\" (UniqueName: \"kubernetes.io/projected/ea124f6f-c7d8-4ff1-8934-d34e18274169-kube-api-access-4jpvb\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.154719 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-config-data\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.257629 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-run-httpd\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.257729 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.257770 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jpvb\" (UniqueName: \"kubernetes.io/projected/ea124f6f-c7d8-4ff1-8934-d34e18274169-kube-api-access-4jpvb\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.257804 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-config-data\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.257893 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-log-httpd\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.257965 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.258006 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-scripts\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.259227 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-run-httpd\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.259308 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-log-httpd\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.270147 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-scripts\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.270365 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-config-data\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.270901 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.275858 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jpvb\" (UniqueName: \"kubernetes.io/projected/ea124f6f-c7d8-4ff1-8934-d34e18274169-kube-api-access-4jpvb\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.276715 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.412735 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.821499 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-px6bb" event={"ID":"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9","Type":"ContainerStarted","Data":"0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd"} Dec 03 00:50:37 crc kubenswrapper[4912]: I1203 00:50:37.853792 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-px6bb" podStartSLOduration=3.397945843 podStartE2EDuration="5.85376626s" podCreationTimestamp="2025-12-03 00:50:32 +0000 UTC" firstStartedPulling="2025-12-03 00:50:34.658208248 +0000 UTC m=+1620.300228808" lastFinishedPulling="2025-12-03 00:50:37.114028665 +0000 UTC m=+1622.756049225" observedRunningTime="2025-12-03 00:50:37.853012991 +0000 UTC m=+1623.495033561" watchObservedRunningTime="2025-12-03 00:50:37.85376626 +0000 UTC m=+1623.495786820" Dec 03 00:50:38 crc kubenswrapper[4912]: I1203 00:50:38.148106 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:38 crc kubenswrapper[4912]: W1203 00:50:38.151524 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea124f6f_c7d8_4ff1_8934_d34e18274169.slice/crio-72c7eaee529b5ef32b928c7eda017878c83bd7709497bd24a533466582b58f19 WatchSource:0}: Error finding container 72c7eaee529b5ef32b928c7eda017878c83bd7709497bd24a533466582b58f19: Status 404 returned error can't find the container with id 72c7eaee529b5ef32b928c7eda017878c83bd7709497bd24a533466582b58f19 Dec 03 00:50:38 crc kubenswrapper[4912]: I1203 00:50:38.586017 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c423153-e5d1-49ca-bba7-70f2f1174226" path="/var/lib/kubelet/pods/2c423153-e5d1-49ca-bba7-70f2f1174226/volumes" Dec 03 00:50:38 crc kubenswrapper[4912]: I1203 00:50:38.849084 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerStarted","Data":"72c7eaee529b5ef32b928c7eda017878c83bd7709497bd24a533466582b58f19"} Dec 03 00:50:39 crc kubenswrapper[4912]: I1203 00:50:39.738400 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:39 crc kubenswrapper[4912]: I1203 00:50:39.862069 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerStarted","Data":"d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb"} Dec 03 00:50:39 crc kubenswrapper[4912]: I1203 00:50:39.862142 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerStarted","Data":"43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd"} Dec 03 00:50:40 crc kubenswrapper[4912]: I1203 00:50:40.876472 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerStarted","Data":"67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1"} Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.575229 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:50:42 crc kubenswrapper[4912]: E1203 00:50:42.576348 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.845692 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.846043 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.900324 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.913901 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="ceilometer-central-agent" containerID="cri-o://43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd" gracePeriod=30 Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.914305 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerStarted","Data":"df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020"} Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.914362 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.914420 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="proxy-httpd" containerID="cri-o://df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020" gracePeriod=30 Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.914506 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="sg-core" containerID="cri-o://67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1" gracePeriod=30 Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.914563 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="ceilometer-notification-agent" containerID="cri-o://d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb" gracePeriod=30 Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.974946 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.14258451 podStartE2EDuration="5.974911418s" podCreationTimestamp="2025-12-03 00:50:37 +0000 UTC" firstStartedPulling="2025-12-03 00:50:38.154478761 +0000 UTC m=+1623.796499321" lastFinishedPulling="2025-12-03 00:50:41.986805669 +0000 UTC m=+1627.628826229" observedRunningTime="2025-12-03 00:50:42.966603076 +0000 UTC m=+1628.608623656" watchObservedRunningTime="2025-12-03 00:50:42.974911418 +0000 UTC m=+1628.616931978" Dec 03 00:50:42 crc kubenswrapper[4912]: I1203 00:50:42.981464 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:43 crc kubenswrapper[4912]: I1203 00:50:43.148177 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-px6bb"] Dec 03 00:50:43 crc kubenswrapper[4912]: I1203 00:50:43.932809 4912 generic.go:334] "Generic (PLEG): container finished" podID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerID="df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020" exitCode=0 Dec 03 00:50:43 crc kubenswrapper[4912]: I1203 00:50:43.933339 4912 generic.go:334] "Generic (PLEG): container finished" podID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerID="67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1" exitCode=2 Dec 03 00:50:43 crc kubenswrapper[4912]: I1203 00:50:43.933353 4912 generic.go:334] "Generic (PLEG): container finished" podID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerID="d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb" exitCode=0 Dec 03 00:50:43 crc kubenswrapper[4912]: I1203 00:50:43.932917 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerDied","Data":"df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020"} Dec 03 00:50:43 crc kubenswrapper[4912]: I1203 00:50:43.933463 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerDied","Data":"67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1"} Dec 03 00:50:43 crc kubenswrapper[4912]: I1203 00:50:43.933490 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerDied","Data":"d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb"} Dec 03 00:50:43 crc kubenswrapper[4912]: I1203 00:50:43.936265 4912 generic.go:334] "Generic (PLEG): container finished" podID="aeac5989-b199-4fcf-8d5e-c49ac6062bfd" containerID="d8d0315b779492e8da1aad0ac6bd679a8db367aa217f9b35baff966ae1e97134" exitCode=0 Dec 03 00:50:43 crc kubenswrapper[4912]: I1203 00:50:43.936553 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" event={"ID":"aeac5989-b199-4fcf-8d5e-c49ac6062bfd","Type":"ContainerDied","Data":"d8d0315b779492e8da1aad0ac6bd679a8db367aa217f9b35baff966ae1e97134"} Dec 03 00:50:44 crc kubenswrapper[4912]: I1203 00:50:44.950047 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-px6bb" podUID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerName="registry-server" containerID="cri-o://0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd" gracePeriod=2 Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.421655 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.574788 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-config-data\") pod \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.574880 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgvwj\" (UniqueName: \"kubernetes.io/projected/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-kube-api-access-zgvwj\") pod \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.575185 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-combined-ca-bundle\") pod \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.575205 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.575333 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-scripts\") pod \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\" (UID: \"aeac5989-b199-4fcf-8d5e-c49ac6062bfd\") " Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.582961 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-scripts" (OuterVolumeSpecName: "scripts") pod "aeac5989-b199-4fcf-8d5e-c49ac6062bfd" (UID: "aeac5989-b199-4fcf-8d5e-c49ac6062bfd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.583724 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-kube-api-access-zgvwj" (OuterVolumeSpecName: "kube-api-access-zgvwj") pod "aeac5989-b199-4fcf-8d5e-c49ac6062bfd" (UID: "aeac5989-b199-4fcf-8d5e-c49ac6062bfd"). InnerVolumeSpecName "kube-api-access-zgvwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.619983 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-config-data" (OuterVolumeSpecName: "config-data") pod "aeac5989-b199-4fcf-8d5e-c49ac6062bfd" (UID: "aeac5989-b199-4fcf-8d5e-c49ac6062bfd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.634046 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aeac5989-b199-4fcf-8d5e-c49ac6062bfd" (UID: "aeac5989-b199-4fcf-8d5e-c49ac6062bfd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.678670 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdkl8\" (UniqueName: \"kubernetes.io/projected/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-kube-api-access-cdkl8\") pod \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.678948 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-utilities\") pod \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.679131 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-catalog-content\") pod \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\" (UID: \"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9\") " Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.680023 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.680059 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgvwj\" (UniqueName: \"kubernetes.io/projected/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-kube-api-access-zgvwj\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.680079 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.680093 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aeac5989-b199-4fcf-8d5e-c49ac6062bfd-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.682935 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-utilities" (OuterVolumeSpecName: "utilities") pod "fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" (UID: "fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.683446 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-kube-api-access-cdkl8" (OuterVolumeSpecName: "kube-api-access-cdkl8") pod "fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" (UID: "fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9"). InnerVolumeSpecName "kube-api-access-cdkl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.731559 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" (UID: "fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.781748 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdkl8\" (UniqueName: \"kubernetes.io/projected/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-kube-api-access-cdkl8\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.781803 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.781817 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.965183 4912 generic.go:334] "Generic (PLEG): container finished" podID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerID="0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd" exitCode=0 Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.965270 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-px6bb" event={"ID":"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9","Type":"ContainerDied","Data":"0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd"} Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.965309 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-px6bb" event={"ID":"fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9","Type":"ContainerDied","Data":"4ee09abc49805ffcc5d7538bef971d3abbf4b97231ea91ecd3d379e98e118019"} Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.965330 4912 scope.go:117] "RemoveContainer" containerID="0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.965273 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-px6bb" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.968667 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" event={"ID":"aeac5989-b199-4fcf-8d5e-c49ac6062bfd","Type":"ContainerDied","Data":"c325707086b379052608ca92ce1fb4bc904c5db2c7e10f44386aaaa09194b05c"} Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.968810 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c325707086b379052608ca92ce1fb4bc904c5db2c7e10f44386aaaa09194b05c" Dec 03 00:50:45 crc kubenswrapper[4912]: I1203 00:50:45.968893 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-wpgr9" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.017267 4912 scope.go:117] "RemoveContainer" containerID="5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.033749 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-px6bb"] Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.045624 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-px6bb"] Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.054790 4912 scope.go:117] "RemoveContainer" containerID="0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.080691 4912 scope.go:117] "RemoveContainer" containerID="0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd" Dec 03 00:50:46 crc kubenswrapper[4912]: E1203 00:50:46.081214 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd\": container with ID starting with 0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd not found: ID does not exist" containerID="0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.081251 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd"} err="failed to get container status \"0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd\": rpc error: code = NotFound desc = could not find container \"0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd\": container with ID starting with 0e0785b5d82176273847465ac2eb7673f1e0528161c3c8ff894fac64232853bd not found: ID does not exist" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.081279 4912 scope.go:117] "RemoveContainer" containerID="5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17" Dec 03 00:50:46 crc kubenswrapper[4912]: E1203 00:50:46.081505 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17\": container with ID starting with 5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17 not found: ID does not exist" containerID="5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.081551 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17"} err="failed to get container status \"5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17\": rpc error: code = NotFound desc = could not find container \"5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17\": container with ID starting with 5c4396a164b86f09e25e04c49bbfb90646885ffb18a01023fba7e6d0f31f8f17 not found: ID does not exist" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.081568 4912 scope.go:117] "RemoveContainer" containerID="0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843" Dec 03 00:50:46 crc kubenswrapper[4912]: E1203 00:50:46.081757 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843\": container with ID starting with 0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843 not found: ID does not exist" containerID="0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.081799 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843"} err="failed to get container status \"0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843\": rpc error: code = NotFound desc = could not find container \"0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843\": container with ID starting with 0aa1be490e42b862d41dffd9cb4fa52108e8325bafe071d67905a6a80e73b843 not found: ID does not exist" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.160744 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 00:50:46 crc kubenswrapper[4912]: E1203 00:50:46.161397 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerName="registry-server" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.161412 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerName="registry-server" Dec 03 00:50:46 crc kubenswrapper[4912]: E1203 00:50:46.161498 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeac5989-b199-4fcf-8d5e-c49ac6062bfd" containerName="nova-cell0-conductor-db-sync" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.161507 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeac5989-b199-4fcf-8d5e-c49ac6062bfd" containerName="nova-cell0-conductor-db-sync" Dec 03 00:50:46 crc kubenswrapper[4912]: E1203 00:50:46.161519 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerName="extract-utilities" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.161525 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerName="extract-utilities" Dec 03 00:50:46 crc kubenswrapper[4912]: E1203 00:50:46.161536 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerName="extract-content" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.161542 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerName="extract-content" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.161829 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeac5989-b199-4fcf-8d5e-c49ac6062bfd" containerName="nova-cell0-conductor-db-sync" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.161845 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" containerName="registry-server" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.163206 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.165627 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-4frvc" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.165915 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.176561 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.308382 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d46d472-d813-4f6d-b58a-8d671244960a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2d46d472-d813-4f6d-b58a-8d671244960a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.308467 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d46d472-d813-4f6d-b58a-8d671244960a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2d46d472-d813-4f6d-b58a-8d671244960a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.308607 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-558ml\" (UniqueName: \"kubernetes.io/projected/2d46d472-d813-4f6d-b58a-8d671244960a-kube-api-access-558ml\") pod \"nova-cell0-conductor-0\" (UID: \"2d46d472-d813-4f6d-b58a-8d671244960a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.411811 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-558ml\" (UniqueName: \"kubernetes.io/projected/2d46d472-d813-4f6d-b58a-8d671244960a-kube-api-access-558ml\") pod \"nova-cell0-conductor-0\" (UID: \"2d46d472-d813-4f6d-b58a-8d671244960a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.412081 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d46d472-d813-4f6d-b58a-8d671244960a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2d46d472-d813-4f6d-b58a-8d671244960a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.412144 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d46d472-d813-4f6d-b58a-8d671244960a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2d46d472-d813-4f6d-b58a-8d671244960a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.416814 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d46d472-d813-4f6d-b58a-8d671244960a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2d46d472-d813-4f6d-b58a-8d671244960a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.417399 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d46d472-d813-4f6d-b58a-8d671244960a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2d46d472-d813-4f6d-b58a-8d671244960a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.433350 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-558ml\" (UniqueName: \"kubernetes.io/projected/2d46d472-d813-4f6d-b58a-8d671244960a-kube-api-access-558ml\") pod \"nova-cell0-conductor-0\" (UID: \"2d46d472-d813-4f6d-b58a-8d671244960a\") " pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.527545 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:46 crc kubenswrapper[4912]: I1203 00:50:46.606113 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9" path="/var/lib/kubelet/pods/fab2bf1c-1cda-4c59-a54b-c24db7ec4ff9/volumes" Dec 03 00:50:47 crc kubenswrapper[4912]: I1203 00:50:47.050125 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 00:50:47 crc kubenswrapper[4912]: W1203 00:50:47.060710 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d46d472_d813_4f6d_b58a_8d671244960a.slice/crio-80fd4c085f1b33889357e272aacd13db4f5e57040c36ae0875bf3b079b4650c9 WatchSource:0}: Error finding container 80fd4c085f1b33889357e272aacd13db4f5e57040c36ae0875bf3b079b4650c9: Status 404 returned error can't find the container with id 80fd4c085f1b33889357e272aacd13db4f5e57040c36ae0875bf3b079b4650c9 Dec 03 00:50:48 crc kubenswrapper[4912]: I1203 00:50:48.004026 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2d46d472-d813-4f6d-b58a-8d671244960a","Type":"ContainerStarted","Data":"86d19346a8a8c46c605bccf68e6d66e43ed2cc5f62138af44ffbbaa973b2735a"} Dec 03 00:50:48 crc kubenswrapper[4912]: I1203 00:50:48.004656 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2d46d472-d813-4f6d-b58a-8d671244960a","Type":"ContainerStarted","Data":"80fd4c085f1b33889357e272aacd13db4f5e57040c36ae0875bf3b079b4650c9"} Dec 03 00:50:48 crc kubenswrapper[4912]: I1203 00:50:48.004703 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:48 crc kubenswrapper[4912]: I1203 00:50:48.047877 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.047851908 podStartE2EDuration="2.047851908s" podCreationTimestamp="2025-12-03 00:50:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:50:48.030424352 +0000 UTC m=+1633.672444932" watchObservedRunningTime="2025-12-03 00:50:48.047851908 +0000 UTC m=+1633.689872478" Dec 03 00:50:48 crc kubenswrapper[4912]: I1203 00:50:48.977792 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.046668 4912 generic.go:334] "Generic (PLEG): container finished" podID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerID="43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd" exitCode=0 Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.048612 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.049645 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerDied","Data":"43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd"} Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.049706 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea124f6f-c7d8-4ff1-8934-d34e18274169","Type":"ContainerDied","Data":"72c7eaee529b5ef32b928c7eda017878c83bd7709497bd24a533466582b58f19"} Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.049744 4912 scope.go:117] "RemoveContainer" containerID="df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.081449 4912 scope.go:117] "RemoveContainer" containerID="67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.095229 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-scripts\") pod \"ea124f6f-c7d8-4ff1-8934-d34e18274169\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.095322 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-sg-core-conf-yaml\") pod \"ea124f6f-c7d8-4ff1-8934-d34e18274169\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.095622 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jpvb\" (UniqueName: \"kubernetes.io/projected/ea124f6f-c7d8-4ff1-8934-d34e18274169-kube-api-access-4jpvb\") pod \"ea124f6f-c7d8-4ff1-8934-d34e18274169\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.095666 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-log-httpd\") pod \"ea124f6f-c7d8-4ff1-8934-d34e18274169\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.095702 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-run-httpd\") pod \"ea124f6f-c7d8-4ff1-8934-d34e18274169\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.095766 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-combined-ca-bundle\") pod \"ea124f6f-c7d8-4ff1-8934-d34e18274169\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.095827 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-config-data\") pod \"ea124f6f-c7d8-4ff1-8934-d34e18274169\" (UID: \"ea124f6f-c7d8-4ff1-8934-d34e18274169\") " Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.097036 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ea124f6f-c7d8-4ff1-8934-d34e18274169" (UID: "ea124f6f-c7d8-4ff1-8934-d34e18274169"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.097422 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ea124f6f-c7d8-4ff1-8934-d34e18274169" (UID: "ea124f6f-c7d8-4ff1-8934-d34e18274169"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.102935 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-scripts" (OuterVolumeSpecName: "scripts") pod "ea124f6f-c7d8-4ff1-8934-d34e18274169" (UID: "ea124f6f-c7d8-4ff1-8934-d34e18274169"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.113841 4912 scope.go:117] "RemoveContainer" containerID="d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.120709 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea124f6f-c7d8-4ff1-8934-d34e18274169-kube-api-access-4jpvb" (OuterVolumeSpecName: "kube-api-access-4jpvb") pod "ea124f6f-c7d8-4ff1-8934-d34e18274169" (UID: "ea124f6f-c7d8-4ff1-8934-d34e18274169"). InnerVolumeSpecName "kube-api-access-4jpvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.129296 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ea124f6f-c7d8-4ff1-8934-d34e18274169" (UID: "ea124f6f-c7d8-4ff1-8934-d34e18274169"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.194882 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea124f6f-c7d8-4ff1-8934-d34e18274169" (UID: "ea124f6f-c7d8-4ff1-8934-d34e18274169"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.199026 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.199074 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.199088 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.199103 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jpvb\" (UniqueName: \"kubernetes.io/projected/ea124f6f-c7d8-4ff1-8934-d34e18274169-kube-api-access-4jpvb\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.199120 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.199133 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea124f6f-c7d8-4ff1-8934-d34e18274169-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.216714 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-config-data" (OuterVolumeSpecName: "config-data") pod "ea124f6f-c7d8-4ff1-8934-d34e18274169" (UID: "ea124f6f-c7d8-4ff1-8934-d34e18274169"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.231728 4912 scope.go:117] "RemoveContainer" containerID="43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.266572 4912 scope.go:117] "RemoveContainer" containerID="df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020" Dec 03 00:50:49 crc kubenswrapper[4912]: E1203 00:50:49.267176 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020\": container with ID starting with df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020 not found: ID does not exist" containerID="df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.267241 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020"} err="failed to get container status \"df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020\": rpc error: code = NotFound desc = could not find container \"df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020\": container with ID starting with df92c4f87554d9f83d0e1e8b493c3ddb6d58aabc6efce38f25714670db578020 not found: ID does not exist" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.267279 4912 scope.go:117] "RemoveContainer" containerID="67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1" Dec 03 00:50:49 crc kubenswrapper[4912]: E1203 00:50:49.268064 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1\": container with ID starting with 67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1 not found: ID does not exist" containerID="67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.268156 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1"} err="failed to get container status \"67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1\": rpc error: code = NotFound desc = could not find container \"67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1\": container with ID starting with 67bffe67f97a0e6229fb58862ac194cb018ab1e07c794ad8947def560334c7b1 not found: ID does not exist" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.268223 4912 scope.go:117] "RemoveContainer" containerID="d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb" Dec 03 00:50:49 crc kubenswrapper[4912]: E1203 00:50:49.269026 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb\": container with ID starting with d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb not found: ID does not exist" containerID="d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.269072 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb"} err="failed to get container status \"d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb\": rpc error: code = NotFound desc = could not find container \"d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb\": container with ID starting with d614784789394588eae4174d5952b37960789bbfddf52d0bc37dee942abda8bb not found: ID does not exist" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.269098 4912 scope.go:117] "RemoveContainer" containerID="43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd" Dec 03 00:50:49 crc kubenswrapper[4912]: E1203 00:50:49.269802 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd\": container with ID starting with 43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd not found: ID does not exist" containerID="43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.269842 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd"} err="failed to get container status \"43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd\": rpc error: code = NotFound desc = could not find container \"43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd\": container with ID starting with 43af15309606b83ed4230778eddd192c1872dca7365928bfd7e1a98ba2be3cdd not found: ID does not exist" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.301952 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea124f6f-c7d8-4ff1-8934-d34e18274169-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.397117 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.410519 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.439114 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:49 crc kubenswrapper[4912]: E1203 00:50:49.439744 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="ceilometer-notification-agent" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.439770 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="ceilometer-notification-agent" Dec 03 00:50:49 crc kubenswrapper[4912]: E1203 00:50:49.439787 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="ceilometer-central-agent" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.439794 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="ceilometer-central-agent" Dec 03 00:50:49 crc kubenswrapper[4912]: E1203 00:50:49.439815 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="sg-core" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.439822 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="sg-core" Dec 03 00:50:49 crc kubenswrapper[4912]: E1203 00:50:49.439839 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="proxy-httpd" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.439844 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="proxy-httpd" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.440084 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="ceilometer-notification-agent" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.440101 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="ceilometer-central-agent" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.440114 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="sg-core" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.440130 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" containerName="proxy-httpd" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.442331 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.446074 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.447076 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.462144 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.611402 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.611596 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-scripts\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.611658 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2gfj\" (UniqueName: \"kubernetes.io/projected/863070c5-4b69-4066-af8f-109b2c9510fb-kube-api-access-l2gfj\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.611697 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-run-httpd\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.612085 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.612181 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-log-httpd\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.612315 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-config-data\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.714560 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-scripts\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.714699 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2gfj\" (UniqueName: \"kubernetes.io/projected/863070c5-4b69-4066-af8f-109b2c9510fb-kube-api-access-l2gfj\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.714748 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-run-httpd\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.714804 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.714833 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-log-httpd\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.714878 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-config-data\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.714937 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.715722 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-log-httpd\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.715892 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-run-httpd\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.721247 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-config-data\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.723489 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-scripts\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.723644 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.723764 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.745962 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2gfj\" (UniqueName: \"kubernetes.io/projected/863070c5-4b69-4066-af8f-109b2c9510fb-kube-api-access-l2gfj\") pod \"ceilometer-0\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " pod="openstack/ceilometer-0" Dec 03 00:50:49 crc kubenswrapper[4912]: I1203 00:50:49.831051 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:50:50 crc kubenswrapper[4912]: W1203 00:50:50.367982 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod863070c5_4b69_4066_af8f_109b2c9510fb.slice/crio-53d15e530a7b9b5c965a91aefae0348bb35a20b6ec171070f44a971477a5754e WatchSource:0}: Error finding container 53d15e530a7b9b5c965a91aefae0348bb35a20b6ec171070f44a971477a5754e: Status 404 returned error can't find the container with id 53d15e530a7b9b5c965a91aefae0348bb35a20b6ec171070f44a971477a5754e Dec 03 00:50:50 crc kubenswrapper[4912]: I1203 00:50:50.380487 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:50 crc kubenswrapper[4912]: I1203 00:50:50.602253 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea124f6f-c7d8-4ff1-8934-d34e18274169" path="/var/lib/kubelet/pods/ea124f6f-c7d8-4ff1-8934-d34e18274169/volumes" Dec 03 00:50:51 crc kubenswrapper[4912]: I1203 00:50:51.084662 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerStarted","Data":"9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf"} Dec 03 00:50:51 crc kubenswrapper[4912]: I1203 00:50:51.085096 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerStarted","Data":"53d15e530a7b9b5c965a91aefae0348bb35a20b6ec171070f44a971477a5754e"} Dec 03 00:50:52 crc kubenswrapper[4912]: I1203 00:50:52.100365 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerStarted","Data":"cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23"} Dec 03 00:50:53 crc kubenswrapper[4912]: I1203 00:50:53.132323 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerStarted","Data":"7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab"} Dec 03 00:50:53 crc kubenswrapper[4912]: I1203 00:50:53.982475 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:50:54 crc kubenswrapper[4912]: I1203 00:50:54.147328 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerStarted","Data":"2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c"} Dec 03 00:50:54 crc kubenswrapper[4912]: I1203 00:50:54.147898 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="ceilometer-central-agent" containerID="cri-o://9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf" gracePeriod=30 Dec 03 00:50:54 crc kubenswrapper[4912]: I1203 00:50:54.148008 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="proxy-httpd" containerID="cri-o://2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c" gracePeriod=30 Dec 03 00:50:54 crc kubenswrapper[4912]: I1203 00:50:54.148050 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="sg-core" containerID="cri-o://7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab" gracePeriod=30 Dec 03 00:50:54 crc kubenswrapper[4912]: I1203 00:50:54.148047 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:50:54 crc kubenswrapper[4912]: I1203 00:50:54.148084 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="ceilometer-notification-agent" containerID="cri-o://cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23" gracePeriod=30 Dec 03 00:50:54 crc kubenswrapper[4912]: I1203 00:50:54.187476 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.75851073 podStartE2EDuration="5.187453625s" podCreationTimestamp="2025-12-03 00:50:49 +0000 UTC" firstStartedPulling="2025-12-03 00:50:50.371580697 +0000 UTC m=+1636.013601257" lastFinishedPulling="2025-12-03 00:50:53.800523582 +0000 UTC m=+1639.442544152" observedRunningTime="2025-12-03 00:50:54.175844395 +0000 UTC m=+1639.817864965" watchObservedRunningTime="2025-12-03 00:50:54.187453625 +0000 UTC m=+1639.829474185" Dec 03 00:50:54 crc kubenswrapper[4912]: I1203 00:50:54.581865 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:50:54 crc kubenswrapper[4912]: E1203 00:50:54.582120 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:50:55 crc kubenswrapper[4912]: I1203 00:50:55.163478 4912 generic.go:334] "Generic (PLEG): container finished" podID="863070c5-4b69-4066-af8f-109b2c9510fb" containerID="7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab" exitCode=2 Dec 03 00:50:55 crc kubenswrapper[4912]: I1203 00:50:55.163524 4912 generic.go:334] "Generic (PLEG): container finished" podID="863070c5-4b69-4066-af8f-109b2c9510fb" containerID="cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23" exitCode=0 Dec 03 00:50:55 crc kubenswrapper[4912]: I1203 00:50:55.163555 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerDied","Data":"7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab"} Dec 03 00:50:55 crc kubenswrapper[4912]: I1203 00:50:55.163590 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerDied","Data":"cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23"} Dec 03 00:50:56 crc kubenswrapper[4912]: I1203 00:50:56.606495 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.084498 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-knjhl"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.086532 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.089916 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.090202 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.116364 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-knjhl"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.240786 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.240930 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-scripts\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.240969 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-config-data\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.240998 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl5cx\" (UniqueName: \"kubernetes.io/projected/cc1d4411-e4c3-4107-90b4-77bd821e61dd-kube-api-access-vl5cx\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.343859 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-scripts\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.343920 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-config-data\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.343956 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl5cx\" (UniqueName: \"kubernetes.io/projected/cc1d4411-e4c3-4107-90b4-77bd821e61dd-kube-api-access-vl5cx\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.344054 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.352987 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-config-data\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.356815 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-scripts\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.366818 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.378174 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl5cx\" (UniqueName: \"kubernetes.io/projected/cc1d4411-e4c3-4107-90b4-77bd821e61dd-kube-api-access-vl5cx\") pod \"nova-cell0-cell-mapping-knjhl\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.406522 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.408519 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.414798 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.417553 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.440577 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.515508 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-k77b2"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.517350 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-k77b2" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.549392 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.549477 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb5cb\" (UniqueName: \"kubernetes.io/projected/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-kube-api-access-pb5cb\") pod \"nova-cell1-novncproxy-0\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.549558 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.614721 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-k77b2"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.648164 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.650418 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.652125 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.652185 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb5cb\" (UniqueName: \"kubernetes.io/projected/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-kube-api-access-pb5cb\") pod \"nova-cell1-novncproxy-0\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.652240 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/435cfdb9-676e-448f-bab9-d39e3cd95afd-operator-scripts\") pod \"aodh-db-create-k77b2\" (UID: \"435cfdb9-676e-448f-bab9-d39e3cd95afd\") " pod="openstack/aodh-db-create-k77b2" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.652278 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.652343 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfg4x\" (UniqueName: \"kubernetes.io/projected/435cfdb9-676e-448f-bab9-d39e3cd95afd-kube-api-access-lfg4x\") pod \"aodh-db-create-k77b2\" (UID: \"435cfdb9-676e-448f-bab9-d39e3cd95afd\") " pod="openstack/aodh-db-create-k77b2" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.659008 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.660016 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.686343 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb5cb\" (UniqueName: \"kubernetes.io/projected/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-kube-api-access-pb5cb\") pod \"nova-cell1-novncproxy-0\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.688205 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.743322 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.776921 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/435cfdb9-676e-448f-bab9-d39e3cd95afd-operator-scripts\") pod \"aodh-db-create-k77b2\" (UID: \"435cfdb9-676e-448f-bab9-d39e3cd95afd\") " pod="openstack/aodh-db-create-k77b2" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.777087 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.777120 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfg4x\" (UniqueName: \"kubernetes.io/projected/435cfdb9-676e-448f-bab9-d39e3cd95afd-kube-api-access-lfg4x\") pod \"aodh-db-create-k77b2\" (UID: \"435cfdb9-676e-448f-bab9-d39e3cd95afd\") " pod="openstack/aodh-db-create-k77b2" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.777224 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70e7336c-9475-4488-bf3f-5768fece3a8f-logs\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.777298 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-config-data\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.777348 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flv4g\" (UniqueName: \"kubernetes.io/projected/70e7336c-9475-4488-bf3f-5768fece3a8f-kube-api-access-flv4g\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.778799 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/435cfdb9-676e-448f-bab9-d39e3cd95afd-operator-scripts\") pod \"aodh-db-create-k77b2\" (UID: \"435cfdb9-676e-448f-bab9-d39e3cd95afd\") " pod="openstack/aodh-db-create-k77b2" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.837235 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfg4x\" (UniqueName: \"kubernetes.io/projected/435cfdb9-676e-448f-bab9-d39e3cd95afd-kube-api-access-lfg4x\") pod \"aodh-db-create-k77b2\" (UID: \"435cfdb9-676e-448f-bab9-d39e3cd95afd\") " pod="openstack/aodh-db-create-k77b2" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.852719 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-820b-account-create-update-khnl7"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.869106 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.875685 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.880402 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.880737 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.880857 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70e7336c-9475-4488-bf3f-5768fece3a8f-logs\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.880920 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-config-data\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.880967 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flv4g\" (UniqueName: \"kubernetes.io/projected/70e7336c-9475-4488-bf3f-5768fece3a8f-kube-api-access-flv4g\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.889460 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.890792 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.891132 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70e7336c-9475-4488-bf3f-5768fece3a8f-logs\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.895320 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-config-data\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.898672 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.929697 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.936100 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flv4g\" (UniqueName: \"kubernetes.io/projected/70e7336c-9475-4488-bf3f-5768fece3a8f-kube-api-access-flv4g\") pod \"nova-api-0\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " pod="openstack/nova-api-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.940355 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.965691 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-820b-account-create-update-khnl7"] Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.984374 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d946f64a-de87-496f-992f-8922b2e061e2-operator-scripts\") pod \"aodh-820b-account-create-update-khnl7\" (UID: \"d946f64a-de87-496f-992f-8922b2e061e2\") " pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.984718 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9cv2\" (UniqueName: \"kubernetes.io/projected/d946f64a-de87-496f-992f-8922b2e061e2-kube-api-access-f9cv2\") pod \"aodh-820b-account-create-update-khnl7\" (UID: \"d946f64a-de87-496f-992f-8922b2e061e2\") " pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.984982 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-config-data\") pod \"nova-scheduler-0\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " pod="openstack/nova-scheduler-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.985360 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " pod="openstack/nova-scheduler-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.985610 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkn8b\" (UniqueName: \"kubernetes.io/projected/35ff0444-86e0-4755-827a-2abbf5d42256-kube-api-access-bkn8b\") pod \"nova-scheduler-0\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " pod="openstack/nova-scheduler-0" Dec 03 00:50:57 crc kubenswrapper[4912]: I1203 00:50:57.995360 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.002233 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.007198 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.054497 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.072483 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-pkxw5"] Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.079792 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.082021 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-k77b2" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.098785 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-pkxw5"] Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.101298 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d946f64a-de87-496f-992f-8922b2e061e2-operator-scripts\") pod \"aodh-820b-account-create-update-khnl7\" (UID: \"d946f64a-de87-496f-992f-8922b2e061e2\") " pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.101506 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9cv2\" (UniqueName: \"kubernetes.io/projected/d946f64a-de87-496f-992f-8922b2e061e2-kube-api-access-f9cv2\") pod \"aodh-820b-account-create-update-khnl7\" (UID: \"d946f64a-de87-496f-992f-8922b2e061e2\") " pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.101657 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-config-data\") pod \"nova-scheduler-0\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " pod="openstack/nova-scheduler-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.101828 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slrbz\" (UniqueName: \"kubernetes.io/projected/241c4c95-26fb-45ca-8a84-f414506c4d61-kube-api-access-slrbz\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.101912 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-config-data\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.101967 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " pod="openstack/nova-scheduler-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.101997 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.102083 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkn8b\" (UniqueName: \"kubernetes.io/projected/35ff0444-86e0-4755-827a-2abbf5d42256-kube-api-access-bkn8b\") pod \"nova-scheduler-0\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " pod="openstack/nova-scheduler-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.102124 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241c4c95-26fb-45ca-8a84-f414506c4d61-logs\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.116095 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d946f64a-de87-496f-992f-8922b2e061e2-operator-scripts\") pod \"aodh-820b-account-create-update-khnl7\" (UID: \"d946f64a-de87-496f-992f-8922b2e061e2\") " pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.127007 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-config-data\") pod \"nova-scheduler-0\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " pod="openstack/nova-scheduler-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.148289 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.150712 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " pod="openstack/nova-scheduler-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.153525 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkn8b\" (UniqueName: \"kubernetes.io/projected/35ff0444-86e0-4755-827a-2abbf5d42256-kube-api-access-bkn8b\") pod \"nova-scheduler-0\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " pod="openstack/nova-scheduler-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.158223 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9cv2\" (UniqueName: \"kubernetes.io/projected/d946f64a-de87-496f-992f-8922b2e061e2-kube-api-access-f9cv2\") pod \"aodh-820b-account-create-update-khnl7\" (UID: \"d946f64a-de87-496f-992f-8922b2e061e2\") " pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.215178 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217516 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217586 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217624 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217652 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-config\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217737 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slrbz\" (UniqueName: \"kubernetes.io/projected/241c4c95-26fb-45ca-8a84-f414506c4d61-kube-api-access-slrbz\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217773 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-config-data\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217802 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217829 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt28k\" (UniqueName: \"kubernetes.io/projected/e08a827e-2f84-4631-9c79-10c82eba397f-kube-api-access-lt28k\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217886 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241c4c95-26fb-45ca-8a84-f414506c4d61-logs\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.217918 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.219746 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241c4c95-26fb-45ca-8a84-f414506c4d61-logs\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.225240 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-config-data\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.232853 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.239713 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slrbz\" (UniqueName: \"kubernetes.io/projected/241c4c95-26fb-45ca-8a84-f414506c4d61-kube-api-access-slrbz\") pod \"nova-metadata-0\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.262194 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.263237 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-knjhl"] Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.320395 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.320480 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.320509 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-config\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.320622 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt28k\" (UniqueName: \"kubernetes.io/projected/e08a827e-2f84-4631-9c79-10c82eba397f-kube-api-access-lt28k\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.320664 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.320712 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.322566 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.323369 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-config\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.323755 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.324100 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.333229 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.342916 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt28k\" (UniqueName: \"kubernetes.io/projected/e08a827e-2f84-4631-9c79-10c82eba397f-kube-api-access-lt28k\") pod \"dnsmasq-dns-568d7fd7cf-pkxw5\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.377299 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.414204 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.569675 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 00:50:58 crc kubenswrapper[4912]: I1203 00:50:58.830077 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-k77b2"] Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.106672 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxxm4"] Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.108499 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.112220 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.113647 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.129453 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxxm4"] Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.156449 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.156906 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htgkr\" (UniqueName: \"kubernetes.io/projected/d7bdcd2a-1372-4f41-a610-74547e28c9de-kube-api-access-htgkr\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.157181 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-scripts\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.157388 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-config-data\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.199661 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.212478 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-820b-account-create-update-khnl7"] Dec 03 00:50:59 crc kubenswrapper[4912]: W1203 00:50:59.213311 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd946f64a_de87_496f_992f_8922b2e061e2.slice/crio-4f8f3b64a493dcfb9d8885a821722736cf9d3338936d390b7376bbb3fc24bea2 WatchSource:0}: Error finding container 4f8f3b64a493dcfb9d8885a821722736cf9d3338936d390b7376bbb3fc24bea2: Status 404 returned error can't find the container with id 4f8f3b64a493dcfb9d8885a821722736cf9d3338936d390b7376bbb3fc24bea2 Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.260739 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-scripts\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.260802 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-config-data\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.260873 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.260902 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htgkr\" (UniqueName: \"kubernetes.io/projected/d7bdcd2a-1372-4f41-a610-74547e28c9de-kube-api-access-htgkr\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.271520 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-scripts\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.277406 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.281950 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-config-data\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.291677 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-knjhl" event={"ID":"cc1d4411-e4c3-4107-90b4-77bd821e61dd","Type":"ContainerStarted","Data":"6d55544f43c33af5471ca5230284ff68432e49975670653a34aebe15b04d759b"} Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.291757 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-knjhl" event={"ID":"cc1d4411-e4c3-4107-90b4-77bd821e61dd","Type":"ContainerStarted","Data":"7fd5d404e7da44fafdbc20864dfc7c17deef8226b9f51f184c2cf960bcd0efa2"} Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.293851 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htgkr\" (UniqueName: \"kubernetes.io/projected/d7bdcd2a-1372-4f41-a610-74547e28c9de-kube-api-access-htgkr\") pod \"nova-cell1-conductor-db-sync-dxxm4\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.306503 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.310012 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70e7336c-9475-4488-bf3f-5768fece3a8f","Type":"ContainerStarted","Data":"aaa03cc7340b8de041be67369158c1014912ebe71d181ee08c6e01296b67f0d4"} Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.334725 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"19d7c201-bd2f-4843-9a2f-eef0d9e547f9","Type":"ContainerStarted","Data":"83ccf98de4e84fbe28bfbe0aee9afcb9867e43be06d52112e79700a0996bd64d"} Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.334814 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-820b-account-create-update-khnl7" event={"ID":"d946f64a-de87-496f-992f-8922b2e061e2","Type":"ContainerStarted","Data":"4f8f3b64a493dcfb9d8885a821722736cf9d3338936d390b7376bbb3fc24bea2"} Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.340833 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-k77b2" event={"ID":"435cfdb9-676e-448f-bab9-d39e3cd95afd","Type":"ContainerStarted","Data":"b94ce68ff5e0303a26a0f97378464cf51d8756cd2a1d9681f8f4f5cfca7449f7"} Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.340895 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-k77b2" event={"ID":"435cfdb9-676e-448f-bab9-d39e3cd95afd","Type":"ContainerStarted","Data":"2f43cbc573ff8343af1aaef8891b56d188613d4448f9a359c3685ee05844a188"} Dec 03 00:50:59 crc kubenswrapper[4912]: W1203 00:50:59.344106 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode08a827e_2f84_4631_9c79_10c82eba397f.slice/crio-3c2c72f764fc32cb693994726a90e40ba8e00b3a8a3a58c1e25d91494668ccf8 WatchSource:0}: Error finding container 3c2c72f764fc32cb693994726a90e40ba8e00b3a8a3a58c1e25d91494668ccf8: Status 404 returned error can't find the container with id 3c2c72f764fc32cb693994726a90e40ba8e00b3a8a3a58c1e25d91494668ccf8 Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.364275 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-pkxw5"] Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.380120 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-knjhl" podStartSLOduration=2.380084322 podStartE2EDuration="2.380084322s" podCreationTimestamp="2025-12-03 00:50:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:50:59.323525211 +0000 UTC m=+1644.965545771" watchObservedRunningTime="2025-12-03 00:50:59.380084322 +0000 UTC m=+1645.022104882" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.431971 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-k77b2" podStartSLOduration=2.431946017 podStartE2EDuration="2.431946017s" podCreationTimestamp="2025-12-03 00:50:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:50:59.373676231 +0000 UTC m=+1645.015696801" watchObservedRunningTime="2025-12-03 00:50:59.431946017 +0000 UTC m=+1645.073966577" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.443887 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:50:59 crc kubenswrapper[4912]: I1203 00:50:59.506832 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.158913 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxxm4"] Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.375620 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dxxm4" event={"ID":"d7bdcd2a-1372-4f41-a610-74547e28c9de","Type":"ContainerStarted","Data":"2b66c65d4b5f21fc9f97cd00a785f9791811ff6219e4f1d47c6d475745de6c97"} Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.380391 4912 generic.go:334] "Generic (PLEG): container finished" podID="e08a827e-2f84-4631-9c79-10c82eba397f" containerID="32e1e4fd1793c50379ca55561fa1c88a2fbca765be6bdf90df800352be7290e5" exitCode=0 Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.380466 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" event={"ID":"e08a827e-2f84-4631-9c79-10c82eba397f","Type":"ContainerDied","Data":"32e1e4fd1793c50379ca55561fa1c88a2fbca765be6bdf90df800352be7290e5"} Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.380489 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" event={"ID":"e08a827e-2f84-4631-9c79-10c82eba397f","Type":"ContainerStarted","Data":"3c2c72f764fc32cb693994726a90e40ba8e00b3a8a3a58c1e25d91494668ccf8"} Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.401735 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241c4c95-26fb-45ca-8a84-f414506c4d61","Type":"ContainerStarted","Data":"df482892241c93c38533a50a5d50b24ed5680b5d7fb759c69d9a276636685f00"} Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.413115 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35ff0444-86e0-4755-827a-2abbf5d42256","Type":"ContainerStarted","Data":"110a83dc053c0c126fdbcf5f689e9150c291ec0bb1158097f480ebc04ee907dd"} Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.430270 4912 generic.go:334] "Generic (PLEG): container finished" podID="d946f64a-de87-496f-992f-8922b2e061e2" containerID="e7fc4eb2633528ebde3b03fb7b350e58a1a0a23594a6494c01b2c7016451eba8" exitCode=0 Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.430363 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-820b-account-create-update-khnl7" event={"ID":"d946f64a-de87-496f-992f-8922b2e061e2","Type":"ContainerDied","Data":"e7fc4eb2633528ebde3b03fb7b350e58a1a0a23594a6494c01b2c7016451eba8"} Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.436816 4912 generic.go:334] "Generic (PLEG): container finished" podID="435cfdb9-676e-448f-bab9-d39e3cd95afd" containerID="b94ce68ff5e0303a26a0f97378464cf51d8756cd2a1d9681f8f4f5cfca7449f7" exitCode=0 Dec 03 00:51:00 crc kubenswrapper[4912]: I1203 00:51:00.437570 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-k77b2" event={"ID":"435cfdb9-676e-448f-bab9-d39e3cd95afd","Type":"ContainerDied","Data":"b94ce68ff5e0303a26a0f97378464cf51d8756cd2a1d9681f8f4f5cfca7449f7"} Dec 03 00:51:01 crc kubenswrapper[4912]: I1203 00:51:01.476828 4912 generic.go:334] "Generic (PLEG): container finished" podID="863070c5-4b69-4066-af8f-109b2c9510fb" containerID="9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf" exitCode=0 Dec 03 00:51:01 crc kubenswrapper[4912]: I1203 00:51:01.477249 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerDied","Data":"9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf"} Dec 03 00:51:01 crc kubenswrapper[4912]: I1203 00:51:01.486312 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dxxm4" event={"ID":"d7bdcd2a-1372-4f41-a610-74547e28c9de","Type":"ContainerStarted","Data":"0c0f4f27ed268c3a0f014812eeaf7bbbbb75ef2670c50aefd3e025b54b275d4d"} Dec 03 00:51:01 crc kubenswrapper[4912]: I1203 00:51:01.496540 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" event={"ID":"e08a827e-2f84-4631-9c79-10c82eba397f","Type":"ContainerStarted","Data":"4d8ef5d001aa2e826b865b2c60bc27c58ada9507f12132f98c5d6bbf394f8042"} Dec 03 00:51:01 crc kubenswrapper[4912]: I1203 00:51:01.496993 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:51:01 crc kubenswrapper[4912]: I1203 00:51:01.512016 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-dxxm4" podStartSLOduration=2.511988997 podStartE2EDuration="2.511988997s" podCreationTimestamp="2025-12-03 00:50:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:01.511062902 +0000 UTC m=+1647.153083472" watchObservedRunningTime="2025-12-03 00:51:01.511988997 +0000 UTC m=+1647.154009547" Dec 03 00:51:01 crc kubenswrapper[4912]: I1203 00:51:01.550339 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" podStartSLOduration=4.550319171 podStartE2EDuration="4.550319171s" podCreationTimestamp="2025-12-03 00:50:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:01.546881239 +0000 UTC m=+1647.188901809" watchObservedRunningTime="2025-12-03 00:51:01.550319171 +0000 UTC m=+1647.192339731" Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.422882 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-k77b2" Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.513519 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/435cfdb9-676e-448f-bab9-d39e3cd95afd-operator-scripts\") pod \"435cfdb9-676e-448f-bab9-d39e3cd95afd\" (UID: \"435cfdb9-676e-448f-bab9-d39e3cd95afd\") " Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.513647 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfg4x\" (UniqueName: \"kubernetes.io/projected/435cfdb9-676e-448f-bab9-d39e3cd95afd-kube-api-access-lfg4x\") pod \"435cfdb9-676e-448f-bab9-d39e3cd95afd\" (UID: \"435cfdb9-676e-448f-bab9-d39e3cd95afd\") " Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.515602 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/435cfdb9-676e-448f-bab9-d39e3cd95afd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "435cfdb9-676e-448f-bab9-d39e3cd95afd" (UID: "435cfdb9-676e-448f-bab9-d39e3cd95afd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.532574 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/435cfdb9-676e-448f-bab9-d39e3cd95afd-kube-api-access-lfg4x" (OuterVolumeSpecName: "kube-api-access-lfg4x") pod "435cfdb9-676e-448f-bab9-d39e3cd95afd" (UID: "435cfdb9-676e-448f-bab9-d39e3cd95afd"). InnerVolumeSpecName "kube-api-access-lfg4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.536327 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-k77b2" Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.536467 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-k77b2" event={"ID":"435cfdb9-676e-448f-bab9-d39e3cd95afd","Type":"ContainerDied","Data":"2f43cbc573ff8343af1aaef8891b56d188613d4448f9a359c3685ee05844a188"} Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.536505 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f43cbc573ff8343af1aaef8891b56d188613d4448f9a359c3685ee05844a188" Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.622915 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/435cfdb9-676e-448f-bab9-d39e3cd95afd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.622956 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfg4x\" (UniqueName: \"kubernetes.io/projected/435cfdb9-676e-448f-bab9-d39e3cd95afd-kube-api-access-lfg4x\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.793750 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 00:51:02 crc kubenswrapper[4912]: I1203 00:51:02.841847 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.044680 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.178066 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9cv2\" (UniqueName: \"kubernetes.io/projected/d946f64a-de87-496f-992f-8922b2e061e2-kube-api-access-f9cv2\") pod \"d946f64a-de87-496f-992f-8922b2e061e2\" (UID: \"d946f64a-de87-496f-992f-8922b2e061e2\") " Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.178609 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d946f64a-de87-496f-992f-8922b2e061e2-operator-scripts\") pod \"d946f64a-de87-496f-992f-8922b2e061e2\" (UID: \"d946f64a-de87-496f-992f-8922b2e061e2\") " Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.181271 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d946f64a-de87-496f-992f-8922b2e061e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d946f64a-de87-496f-992f-8922b2e061e2" (UID: "d946f64a-de87-496f-992f-8922b2e061e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.193977 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d946f64a-de87-496f-992f-8922b2e061e2-kube-api-access-f9cv2" (OuterVolumeSpecName: "kube-api-access-f9cv2") pod "d946f64a-de87-496f-992f-8922b2e061e2" (UID: "d946f64a-de87-496f-992f-8922b2e061e2"). InnerVolumeSpecName "kube-api-access-f9cv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.282055 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d946f64a-de87-496f-992f-8922b2e061e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.282286 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9cv2\" (UniqueName: \"kubernetes.io/projected/d946f64a-de87-496f-992f-8922b2e061e2-kube-api-access-f9cv2\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.567359 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241c4c95-26fb-45ca-8a84-f414506c4d61","Type":"ContainerStarted","Data":"220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3"} Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.567870 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241c4c95-26fb-45ca-8a84-f414506c4d61","Type":"ContainerStarted","Data":"e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75"} Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.567586 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerName="nova-metadata-log" containerID="cri-o://e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75" gracePeriod=30 Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.567628 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerName="nova-metadata-metadata" containerID="cri-o://220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3" gracePeriod=30 Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.595321 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="19d7c201-bd2f-4843-9a2f-eef0d9e547f9" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67" gracePeriod=30 Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.596091 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70e7336c-9475-4488-bf3f-5768fece3a8f","Type":"ContainerStarted","Data":"bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b"} Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.596133 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70e7336c-9475-4488-bf3f-5768fece3a8f","Type":"ContainerStarted","Data":"0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f"} Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.596143 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35ff0444-86e0-4755-827a-2abbf5d42256","Type":"ContainerStarted","Data":"ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe"} Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.596156 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"19d7c201-bd2f-4843-9a2f-eef0d9e547f9","Type":"ContainerStarted","Data":"d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67"} Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.620484 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-820b-account-create-update-khnl7" event={"ID":"d946f64a-de87-496f-992f-8922b2e061e2","Type":"ContainerDied","Data":"4f8f3b64a493dcfb9d8885a821722736cf9d3338936d390b7376bbb3fc24bea2"} Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.620541 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f8f3b64a493dcfb9d8885a821722736cf9d3338936d390b7376bbb3fc24bea2" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.620633 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-820b-account-create-update-khnl7" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.623456 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.420185809 podStartE2EDuration="7.623400092s" podCreationTimestamp="2025-12-03 00:50:57 +0000 UTC" firstStartedPulling="2025-12-03 00:50:59.534180527 +0000 UTC m=+1645.176201087" lastFinishedPulling="2025-12-03 00:51:03.73739481 +0000 UTC m=+1649.379415370" observedRunningTime="2025-12-03 00:51:04.598926048 +0000 UTC m=+1650.240946618" watchObservedRunningTime="2025-12-03 00:51:04.623400092 +0000 UTC m=+1650.265420652" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.647575 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5448041310000002 podStartE2EDuration="7.647556517s" podCreationTimestamp="2025-12-03 00:50:57 +0000 UTC" firstStartedPulling="2025-12-03 00:50:58.635277861 +0000 UTC m=+1644.277298421" lastFinishedPulling="2025-12-03 00:51:03.738030247 +0000 UTC m=+1649.380050807" observedRunningTime="2025-12-03 00:51:04.64616035 +0000 UTC m=+1650.288180940" watchObservedRunningTime="2025-12-03 00:51:04.647556517 +0000 UTC m=+1650.289577067" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.674491 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.229387584 podStartE2EDuration="7.674467756s" podCreationTimestamp="2025-12-03 00:50:57 +0000 UTC" firstStartedPulling="2025-12-03 00:50:59.29203048 +0000 UTC m=+1644.934051040" lastFinishedPulling="2025-12-03 00:51:03.737110652 +0000 UTC m=+1649.379131212" observedRunningTime="2025-12-03 00:51:04.669095262 +0000 UTC m=+1650.311115822" watchObservedRunningTime="2025-12-03 00:51:04.674467756 +0000 UTC m=+1650.316488316" Dec 03 00:51:04 crc kubenswrapper[4912]: I1203 00:51:04.704377 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.176175633 podStartE2EDuration="7.704350134s" podCreationTimestamp="2025-12-03 00:50:57 +0000 UTC" firstStartedPulling="2025-12-03 00:50:59.207906224 +0000 UTC m=+1644.849926774" lastFinishedPulling="2025-12-03 00:51:03.736080715 +0000 UTC m=+1649.378101275" observedRunningTime="2025-12-03 00:51:04.691341696 +0000 UTC m=+1650.333362276" watchObservedRunningTime="2025-12-03 00:51:04.704350134 +0000 UTC m=+1650.346370694" Dec 03 00:51:05 crc kubenswrapper[4912]: I1203 00:51:05.635697 4912 generic.go:334] "Generic (PLEG): container finished" podID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerID="e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75" exitCode=143 Dec 03 00:51:05 crc kubenswrapper[4912]: I1203 00:51:05.635952 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241c4c95-26fb-45ca-8a84-f414506c4d61","Type":"ContainerDied","Data":"e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75"} Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.931320 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.972799 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-z6bqt"] Dec 03 00:51:07 crc kubenswrapper[4912]: E1203 00:51:07.973560 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d946f64a-de87-496f-992f-8922b2e061e2" containerName="mariadb-account-create-update" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.973582 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d946f64a-de87-496f-992f-8922b2e061e2" containerName="mariadb-account-create-update" Dec 03 00:51:07 crc kubenswrapper[4912]: E1203 00:51:07.973602 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435cfdb9-676e-448f-bab9-d39e3cd95afd" containerName="mariadb-database-create" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.973614 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="435cfdb9-676e-448f-bab9-d39e3cd95afd" containerName="mariadb-database-create" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.973894 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="435cfdb9-676e-448f-bab9-d39e3cd95afd" containerName="mariadb-database-create" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.973942 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d946f64a-de87-496f-992f-8922b2e061e2" containerName="mariadb-account-create-update" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.975079 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.977631 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.977773 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.977985 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.978205 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tnsxp" Dec 03 00:51:07 crc kubenswrapper[4912]: I1203 00:51:07.990657 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-z6bqt"] Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.135396 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-scripts\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.135899 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r5rn\" (UniqueName: \"kubernetes.io/projected/cfd2dfcd-2539-425b-9071-d0de6bc9a598-kube-api-access-8r5rn\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.135937 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-config-data\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.137207 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-combined-ca-bundle\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.149079 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.149138 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.239790 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-scripts\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.239916 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r5rn\" (UniqueName: \"kubernetes.io/projected/cfd2dfcd-2539-425b-9071-d0de6bc9a598-kube-api-access-8r5rn\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.239954 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-config-data\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.240056 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-combined-ca-bundle\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.250728 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-combined-ca-bundle\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.252116 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-config-data\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.263157 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.263469 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.269198 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r5rn\" (UniqueName: \"kubernetes.io/projected/cfd2dfcd-2539-425b-9071-d0de6bc9a598-kube-api-access-8r5rn\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.278226 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-scripts\") pod \"aodh-db-sync-z6bqt\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.306487 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.318547 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.377686 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.377785 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.430877 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.575873 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:51:08 crc kubenswrapper[4912]: E1203 00:51:08.576397 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.630774 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-g7nr8"] Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.632600 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" podUID="dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" containerName="dnsmasq-dns" containerID="cri-o://b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea" gracePeriod=10 Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.743197 4912 generic.go:334] "Generic (PLEG): container finished" podID="cc1d4411-e4c3-4107-90b4-77bd821e61dd" containerID="6d55544f43c33af5471ca5230284ff68432e49975670653a34aebe15b04d759b" exitCode=0 Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.743817 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-knjhl" event={"ID":"cc1d4411-e4c3-4107-90b4-77bd821e61dd","Type":"ContainerDied","Data":"6d55544f43c33af5471ca5230284ff68432e49975670653a34aebe15b04d759b"} Dec 03 00:51:08 crc kubenswrapper[4912]: I1203 00:51:08.807976 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.052050 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-z6bqt"] Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.235728 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.219:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.235825 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.219:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.384070 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.424845 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-nb\") pod \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.424968 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-svc\") pod \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.425104 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-sb\") pod \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.425237 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-swift-storage-0\") pod \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.425282 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-config\") pod \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.425331 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44zd9\" (UniqueName: \"kubernetes.io/projected/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-kube-api-access-44zd9\") pod \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\" (UID: \"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02\") " Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.442610 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-kube-api-access-44zd9" (OuterVolumeSpecName: "kube-api-access-44zd9") pod "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" (UID: "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02"). InnerVolumeSpecName "kube-api-access-44zd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.521100 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" (UID: "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.522549 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" (UID: "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.529980 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.530017 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.530030 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44zd9\" (UniqueName: \"kubernetes.io/projected/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-kube-api-access-44zd9\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.543150 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" (UID: "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.543192 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" (UID: "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.566066 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-config" (OuterVolumeSpecName: "config") pod "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" (UID: "dede53be-0800-4bd0-a6ee-2c9e2b9c9d02"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.633503 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.633564 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.633579 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.762337 4912 generic.go:334] "Generic (PLEG): container finished" podID="dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" containerID="b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea" exitCode=0 Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.762410 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" event={"ID":"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02","Type":"ContainerDied","Data":"b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea"} Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.762463 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" event={"ID":"dede53be-0800-4bd0-a6ee-2c9e2b9c9d02","Type":"ContainerDied","Data":"0bba508da0571926191f604728f757b5b98f2dd18da87dd797dfbf15e6c76dd9"} Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.762485 4912 scope.go:117] "RemoveContainer" containerID="b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.762627 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-g7nr8" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.768411 4912 generic.go:334] "Generic (PLEG): container finished" podID="d7bdcd2a-1372-4f41-a610-74547e28c9de" containerID="0c0f4f27ed268c3a0f014812eeaf7bbbbb75ef2670c50aefd3e025b54b275d4d" exitCode=0 Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.768514 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dxxm4" event={"ID":"d7bdcd2a-1372-4f41-a610-74547e28c9de","Type":"ContainerDied","Data":"0c0f4f27ed268c3a0f014812eeaf7bbbbb75ef2670c50aefd3e025b54b275d4d"} Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.777290 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-z6bqt" event={"ID":"cfd2dfcd-2539-425b-9071-d0de6bc9a598","Type":"ContainerStarted","Data":"0b1985743d6271b8708817b58d8e54b04f189a416b56a81379bc938e2420789e"} Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.832277 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-g7nr8"] Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.840287 4912 scope.go:117] "RemoveContainer" containerID="71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.847716 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-g7nr8"] Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.874822 4912 scope.go:117] "RemoveContainer" containerID="b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea" Dec 03 00:51:09 crc kubenswrapper[4912]: E1203 00:51:09.875569 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea\": container with ID starting with b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea not found: ID does not exist" containerID="b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.875647 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea"} err="failed to get container status \"b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea\": rpc error: code = NotFound desc = could not find container \"b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea\": container with ID starting with b09e5c4fcf33f56091d99ab10c06ad185a4160b3a62a2d2d8cad202977e6cfea not found: ID does not exist" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.875713 4912 scope.go:117] "RemoveContainer" containerID="71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f" Dec 03 00:51:09 crc kubenswrapper[4912]: E1203 00:51:09.876133 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f\": container with ID starting with 71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f not found: ID does not exist" containerID="71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f" Dec 03 00:51:09 crc kubenswrapper[4912]: I1203 00:51:09.876173 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f"} err="failed to get container status \"71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f\": rpc error: code = NotFound desc = could not find container \"71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f\": container with ID starting with 71f947ff707c40c03651bc9e50626b09ee2a56fb1d9962ef0bd6cc4a6d0bd21f not found: ID does not exist" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.450836 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.477111 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-combined-ca-bundle\") pod \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.477264 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl5cx\" (UniqueName: \"kubernetes.io/projected/cc1d4411-e4c3-4107-90b4-77bd821e61dd-kube-api-access-vl5cx\") pod \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.477374 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-config-data\") pod \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.477423 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-scripts\") pod \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\" (UID: \"cc1d4411-e4c3-4107-90b4-77bd821e61dd\") " Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.521687 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc1d4411-e4c3-4107-90b4-77bd821e61dd-kube-api-access-vl5cx" (OuterVolumeSpecName: "kube-api-access-vl5cx") pod "cc1d4411-e4c3-4107-90b4-77bd821e61dd" (UID: "cc1d4411-e4c3-4107-90b4-77bd821e61dd"). InnerVolumeSpecName "kube-api-access-vl5cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.530732 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-scripts" (OuterVolumeSpecName: "scripts") pod "cc1d4411-e4c3-4107-90b4-77bd821e61dd" (UID: "cc1d4411-e4c3-4107-90b4-77bd821e61dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.582186 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl5cx\" (UniqueName: \"kubernetes.io/projected/cc1d4411-e4c3-4107-90b4-77bd821e61dd-kube-api-access-vl5cx\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.582232 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.590830 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc1d4411-e4c3-4107-90b4-77bd821e61dd" (UID: "cc1d4411-e4c3-4107-90b4-77bd821e61dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.592342 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" path="/var/lib/kubelet/pods/dede53be-0800-4bd0-a6ee-2c9e2b9c9d02/volumes" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.602569 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-config-data" (OuterVolumeSpecName: "config-data") pod "cc1d4411-e4c3-4107-90b4-77bd821e61dd" (UID: "cc1d4411-e4c3-4107-90b4-77bd821e61dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.686479 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.686540 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc1d4411-e4c3-4107-90b4-77bd821e61dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.824563 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-knjhl" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.825372 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-knjhl" event={"ID":"cc1d4411-e4c3-4107-90b4-77bd821e61dd","Type":"ContainerDied","Data":"7fd5d404e7da44fafdbc20864dfc7c17deef8226b9f51f184c2cf960bcd0efa2"} Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.825525 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fd5d404e7da44fafdbc20864dfc7c17deef8226b9f51f184c2cf960bcd0efa2" Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.977560 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.978389 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-log" containerID="cri-o://0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f" gracePeriod=30 Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.978803 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-api" containerID="cri-o://bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b" gracePeriod=30 Dec 03 00:51:10 crc kubenswrapper[4912]: I1203 00:51:10.985001 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.221887 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.305974 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-scripts\") pod \"d7bdcd2a-1372-4f41-a610-74547e28c9de\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.306596 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-config-data\") pod \"d7bdcd2a-1372-4f41-a610-74547e28c9de\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.306776 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-combined-ca-bundle\") pod \"d7bdcd2a-1372-4f41-a610-74547e28c9de\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.306878 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htgkr\" (UniqueName: \"kubernetes.io/projected/d7bdcd2a-1372-4f41-a610-74547e28c9de-kube-api-access-htgkr\") pod \"d7bdcd2a-1372-4f41-a610-74547e28c9de\" (UID: \"d7bdcd2a-1372-4f41-a610-74547e28c9de\") " Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.313398 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-scripts" (OuterVolumeSpecName: "scripts") pod "d7bdcd2a-1372-4f41-a610-74547e28c9de" (UID: "d7bdcd2a-1372-4f41-a610-74547e28c9de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.317704 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7bdcd2a-1372-4f41-a610-74547e28c9de-kube-api-access-htgkr" (OuterVolumeSpecName: "kube-api-access-htgkr") pod "d7bdcd2a-1372-4f41-a610-74547e28c9de" (UID: "d7bdcd2a-1372-4f41-a610-74547e28c9de"). InnerVolumeSpecName "kube-api-access-htgkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.348936 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-config-data" (OuterVolumeSpecName: "config-data") pod "d7bdcd2a-1372-4f41-a610-74547e28c9de" (UID: "d7bdcd2a-1372-4f41-a610-74547e28c9de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.362198 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7bdcd2a-1372-4f41-a610-74547e28c9de" (UID: "d7bdcd2a-1372-4f41-a610-74547e28c9de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.410034 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.410081 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.410095 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htgkr\" (UniqueName: \"kubernetes.io/projected/d7bdcd2a-1372-4f41-a610-74547e28c9de-kube-api-access-htgkr\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.410111 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7bdcd2a-1372-4f41-a610-74547e28c9de-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.856045 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dxxm4" event={"ID":"d7bdcd2a-1372-4f41-a610-74547e28c9de","Type":"ContainerDied","Data":"2b66c65d4b5f21fc9f97cd00a785f9791811ff6219e4f1d47c6d475745de6c97"} Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.856107 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b66c65d4b5f21fc9f97cd00a785f9791811ff6219e4f1d47c6d475745de6c97" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.856261 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dxxm4" Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.860728 4912 generic.go:334] "Generic (PLEG): container finished" podID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerID="0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f" exitCode=143 Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.861108 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="35ff0444-86e0-4755-827a-2abbf5d42256" containerName="nova-scheduler-scheduler" containerID="cri-o://ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe" gracePeriod=30 Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.861268 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70e7336c-9475-4488-bf3f-5768fece3a8f","Type":"ContainerDied","Data":"0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f"} Dec 03 00:51:11 crc kubenswrapper[4912]: I1203 00:51:11.910969 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 00:51:12 crc kubenswrapper[4912]: E1203 00:51:12.008201 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" containerName="init" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.008259 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" containerName="init" Dec 03 00:51:12 crc kubenswrapper[4912]: E1203 00:51:12.008350 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" containerName="dnsmasq-dns" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.008359 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" containerName="dnsmasq-dns" Dec 03 00:51:12 crc kubenswrapper[4912]: E1203 00:51:12.008439 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7bdcd2a-1372-4f41-a610-74547e28c9de" containerName="nova-cell1-conductor-db-sync" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.014564 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7bdcd2a-1372-4f41-a610-74547e28c9de" containerName="nova-cell1-conductor-db-sync" Dec 03 00:51:12 crc kubenswrapper[4912]: E1203 00:51:12.014611 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1d4411-e4c3-4107-90b4-77bd821e61dd" containerName="nova-manage" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.014619 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1d4411-e4c3-4107-90b4-77bd821e61dd" containerName="nova-manage" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.016095 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc1d4411-e4c3-4107-90b4-77bd821e61dd" containerName="nova-manage" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.016168 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="dede53be-0800-4bd0-a6ee-2c9e2b9c9d02" containerName="dnsmasq-dns" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.016195 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7bdcd2a-1372-4f41-a610-74547e28c9de" containerName="nova-cell1-conductor-db-sync" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.041641 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.041770 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.054824 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.209605 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8116ee8d-aa67-4a8f-a765-b846f397f7d4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8116ee8d-aa67-4a8f-a765-b846f397f7d4\") " pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.209706 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qtnq\" (UniqueName: \"kubernetes.io/projected/8116ee8d-aa67-4a8f-a765-b846f397f7d4-kube-api-access-8qtnq\") pod \"nova-cell1-conductor-0\" (UID: \"8116ee8d-aa67-4a8f-a765-b846f397f7d4\") " pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.210394 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8116ee8d-aa67-4a8f-a765-b846f397f7d4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8116ee8d-aa67-4a8f-a765-b846f397f7d4\") " pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.312569 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8116ee8d-aa67-4a8f-a765-b846f397f7d4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8116ee8d-aa67-4a8f-a765-b846f397f7d4\") " pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.312658 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qtnq\" (UniqueName: \"kubernetes.io/projected/8116ee8d-aa67-4a8f-a765-b846f397f7d4-kube-api-access-8qtnq\") pod \"nova-cell1-conductor-0\" (UID: \"8116ee8d-aa67-4a8f-a765-b846f397f7d4\") " pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.312812 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8116ee8d-aa67-4a8f-a765-b846f397f7d4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8116ee8d-aa67-4a8f-a765-b846f397f7d4\") " pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.319760 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8116ee8d-aa67-4a8f-a765-b846f397f7d4-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8116ee8d-aa67-4a8f-a765-b846f397f7d4\") " pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.321685 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8116ee8d-aa67-4a8f-a765-b846f397f7d4-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8116ee8d-aa67-4a8f-a765-b846f397f7d4\") " pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.335974 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qtnq\" (UniqueName: \"kubernetes.io/projected/8116ee8d-aa67-4a8f-a765-b846f397f7d4-kube-api-access-8qtnq\") pod \"nova-cell1-conductor-0\" (UID: \"8116ee8d-aa67-4a8f-a765-b846f397f7d4\") " pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:12 crc kubenswrapper[4912]: I1203 00:51:12.387320 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:13 crc kubenswrapper[4912]: E1203 00:51:13.267761 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 00:51:13 crc kubenswrapper[4912]: E1203 00:51:13.271322 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 00:51:13 crc kubenswrapper[4912]: E1203 00:51:13.273324 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 00:51:13 crc kubenswrapper[4912]: E1203 00:51:13.273376 4912 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="35ff0444-86e0-4755-827a-2abbf5d42256" containerName="nova-scheduler-scheduler" Dec 03 00:51:15 crc kubenswrapper[4912]: I1203 00:51:15.715618 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 00:51:15 crc kubenswrapper[4912]: I1203 00:51:15.933886 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-z6bqt" event={"ID":"cfd2dfcd-2539-425b-9071-d0de6bc9a598","Type":"ContainerStarted","Data":"9b9e8f9697d2a56d7564f1fd10315324cd057bf556cf51049d076bf7644a6aa3"} Dec 03 00:51:15 crc kubenswrapper[4912]: I1203 00:51:15.940141 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8116ee8d-aa67-4a8f-a765-b846f397f7d4","Type":"ContainerStarted","Data":"dd868c744ea9cd8b13c41f9c0ff079ad3bfa2d4ba076a86f11d69ddfc810636b"} Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.613697 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.660458 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-z6bqt" podStartSLOduration=3.617982585 podStartE2EDuration="9.660404157s" podCreationTimestamp="2025-12-03 00:51:07 +0000 UTC" firstStartedPulling="2025-12-03 00:51:09.11513585 +0000 UTC m=+1654.757156410" lastFinishedPulling="2025-12-03 00:51:15.157557422 +0000 UTC m=+1660.799577982" observedRunningTime="2025-12-03 00:51:15.963345191 +0000 UTC m=+1661.605365751" watchObservedRunningTime="2025-12-03 00:51:16.660404157 +0000 UTC m=+1662.302424737" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.686999 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkn8b\" (UniqueName: \"kubernetes.io/projected/35ff0444-86e0-4755-827a-2abbf5d42256-kube-api-access-bkn8b\") pod \"35ff0444-86e0-4755-827a-2abbf5d42256\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.687207 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-combined-ca-bundle\") pod \"35ff0444-86e0-4755-827a-2abbf5d42256\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.687387 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-config-data\") pod \"35ff0444-86e0-4755-827a-2abbf5d42256\" (UID: \"35ff0444-86e0-4755-827a-2abbf5d42256\") " Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.707606 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35ff0444-86e0-4755-827a-2abbf5d42256-kube-api-access-bkn8b" (OuterVolumeSpecName: "kube-api-access-bkn8b") pod "35ff0444-86e0-4755-827a-2abbf5d42256" (UID: "35ff0444-86e0-4755-827a-2abbf5d42256"). InnerVolumeSpecName "kube-api-access-bkn8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.747816 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35ff0444-86e0-4755-827a-2abbf5d42256" (UID: "35ff0444-86e0-4755-827a-2abbf5d42256"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.796925 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkn8b\" (UniqueName: \"kubernetes.io/projected/35ff0444-86e0-4755-827a-2abbf5d42256-kube-api-access-bkn8b\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.796965 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.799642 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-config-data" (OuterVolumeSpecName: "config-data") pod "35ff0444-86e0-4755-827a-2abbf5d42256" (UID: "35ff0444-86e0-4755-827a-2abbf5d42256"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.851745 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.900454 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35ff0444-86e0-4755-827a-2abbf5d42256-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.954896 4912 generic.go:334] "Generic (PLEG): container finished" podID="35ff0444-86e0-4755-827a-2abbf5d42256" containerID="ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe" exitCode=0 Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.954958 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35ff0444-86e0-4755-827a-2abbf5d42256","Type":"ContainerDied","Data":"ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe"} Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.955004 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.955046 4912 scope.go:117] "RemoveContainer" containerID="ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.955024 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"35ff0444-86e0-4755-827a-2abbf5d42256","Type":"ContainerDied","Data":"110a83dc053c0c126fdbcf5f689e9150c291ec0bb1158097f480ebc04ee907dd"} Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.959974 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8116ee8d-aa67-4a8f-a765-b846f397f7d4","Type":"ContainerStarted","Data":"6626281e9bad382986a4dae5990bc5cd779bfc357ec90fc870f78cae89d1cd1f"} Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.960127 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.965095 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.965462 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70e7336c-9475-4488-bf3f-5768fece3a8f","Type":"ContainerDied","Data":"bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b"} Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.970614 4912 generic.go:334] "Generic (PLEG): container finished" podID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerID="bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b" exitCode=0 Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.970698 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70e7336c-9475-4488-bf3f-5768fece3a8f","Type":"ContainerDied","Data":"aaa03cc7340b8de041be67369158c1014912ebe71d181ee08c6e01296b67f0d4"} Dec 03 00:51:16 crc kubenswrapper[4912]: I1203 00:51:16.985079 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=5.985055337 podStartE2EDuration="5.985055337s" podCreationTimestamp="2025-12-03 00:51:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:16.983099805 +0000 UTC m=+1662.625120375" watchObservedRunningTime="2025-12-03 00:51:16.985055337 +0000 UTC m=+1662.627075887" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.003171 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70e7336c-9475-4488-bf3f-5768fece3a8f-logs\") pod \"70e7336c-9475-4488-bf3f-5768fece3a8f\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.005383 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70e7336c-9475-4488-bf3f-5768fece3a8f-logs" (OuterVolumeSpecName: "logs") pod "70e7336c-9475-4488-bf3f-5768fece3a8f" (UID: "70e7336c-9475-4488-bf3f-5768fece3a8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.010889 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-combined-ca-bundle\") pod \"70e7336c-9475-4488-bf3f-5768fece3a8f\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.011167 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-config-data\") pod \"70e7336c-9475-4488-bf3f-5768fece3a8f\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.011210 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flv4g\" (UniqueName: \"kubernetes.io/projected/70e7336c-9475-4488-bf3f-5768fece3a8f-kube-api-access-flv4g\") pod \"70e7336c-9475-4488-bf3f-5768fece3a8f\" (UID: \"70e7336c-9475-4488-bf3f-5768fece3a8f\") " Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.012451 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70e7336c-9475-4488-bf3f-5768fece3a8f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.015665 4912 scope.go:117] "RemoveContainer" containerID="ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.016003 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70e7336c-9475-4488-bf3f-5768fece3a8f-kube-api-access-flv4g" (OuterVolumeSpecName: "kube-api-access-flv4g") pod "70e7336c-9475-4488-bf3f-5768fece3a8f" (UID: "70e7336c-9475-4488-bf3f-5768fece3a8f"). InnerVolumeSpecName "kube-api-access-flv4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:17 crc kubenswrapper[4912]: E1203 00:51:17.017096 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe\": container with ID starting with ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe not found: ID does not exist" containerID="ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.017136 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe"} err="failed to get container status \"ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe\": rpc error: code = NotFound desc = could not find container \"ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe\": container with ID starting with ad6eb8a7246eb728894634710d6a8850a2f3a79e85611794913620e66788d1fe not found: ID does not exist" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.017162 4912 scope.go:117] "RemoveContainer" containerID="bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.027986 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.037720 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.054791 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:51:17 crc kubenswrapper[4912]: E1203 00:51:17.055419 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ff0444-86e0-4755-827a-2abbf5d42256" containerName="nova-scheduler-scheduler" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.055462 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ff0444-86e0-4755-827a-2abbf5d42256" containerName="nova-scheduler-scheduler" Dec 03 00:51:17 crc kubenswrapper[4912]: E1203 00:51:17.055503 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-api" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.055512 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-api" Dec 03 00:51:17 crc kubenswrapper[4912]: E1203 00:51:17.055543 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-log" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.055550 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-log" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.055782 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-log" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.055800 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="35ff0444-86e0-4755-827a-2abbf5d42256" containerName="nova-scheduler-scheduler" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.055811 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" containerName="nova-api-api" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.056764 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.059857 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.070761 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70e7336c-9475-4488-bf3f-5768fece3a8f" (UID: "70e7336c-9475-4488-bf3f-5768fece3a8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.075288 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-config-data" (OuterVolumeSpecName: "config-data") pod "70e7336c-9475-4488-bf3f-5768fece3a8f" (UID: "70e7336c-9475-4488-bf3f-5768fece3a8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.078578 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.089247 4912 scope.go:117] "RemoveContainer" containerID="0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.115342 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.115376 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70e7336c-9475-4488-bf3f-5768fece3a8f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.115388 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flv4g\" (UniqueName: \"kubernetes.io/projected/70e7336c-9475-4488-bf3f-5768fece3a8f-kube-api-access-flv4g\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.141005 4912 scope.go:117] "RemoveContainer" containerID="bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b" Dec 03 00:51:17 crc kubenswrapper[4912]: E1203 00:51:17.141688 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b\": container with ID starting with bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b not found: ID does not exist" containerID="bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.141724 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b"} err="failed to get container status \"bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b\": rpc error: code = NotFound desc = could not find container \"bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b\": container with ID starting with bc9497d97c46db9296c9af2e80fca6ecdd73bbc77aa6523bf1a63261d8b01c8b not found: ID does not exist" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.141777 4912 scope.go:117] "RemoveContainer" containerID="0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f" Dec 03 00:51:17 crc kubenswrapper[4912]: E1203 00:51:17.142239 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f\": container with ID starting with 0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f not found: ID does not exist" containerID="0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.142268 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f"} err="failed to get container status \"0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f\": rpc error: code = NotFound desc = could not find container \"0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f\": container with ID starting with 0b35b0e6c06eb66032e1c62b3c8c42a490f59e6ed7e2a5a8f9137c017f46f57f not found: ID does not exist" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.217553 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7twk\" (UniqueName: \"kubernetes.io/projected/8105f90c-0397-4120-b23d-a0d3d72f6b01-kube-api-access-n7twk\") pod \"nova-scheduler-0\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.218188 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-config-data\") pod \"nova-scheduler-0\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.218238 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.322889 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7twk\" (UniqueName: \"kubernetes.io/projected/8105f90c-0397-4120-b23d-a0d3d72f6b01-kube-api-access-n7twk\") pod \"nova-scheduler-0\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.323030 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-config-data\") pod \"nova-scheduler-0\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.323088 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.330569 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.362045 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.365067 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-config-data\") pod \"nova-scheduler-0\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.367315 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7twk\" (UniqueName: \"kubernetes.io/projected/8105f90c-0397-4120-b23d-a0d3d72f6b01-kube-api-access-n7twk\") pod \"nova-scheduler-0\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.390523 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.410067 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.417447 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.420866 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.423829 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.429676 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.528707 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-config-data\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.528777 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80f6c2b-8c99-4863-80de-41146440cf57-logs\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.528829 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twt4c\" (UniqueName: \"kubernetes.io/projected/b80f6c2b-8c99-4863-80de-41146440cf57-kube-api-access-twt4c\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.529072 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.637017 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-config-data\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.637490 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80f6c2b-8c99-4863-80de-41146440cf57-logs\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.637551 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twt4c\" (UniqueName: \"kubernetes.io/projected/b80f6c2b-8c99-4863-80de-41146440cf57-kube-api-access-twt4c\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.637671 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.638339 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80f6c2b-8c99-4863-80de-41146440cf57-logs\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.645975 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.647233 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-config-data\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.654528 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twt4c\" (UniqueName: \"kubernetes.io/projected/b80f6c2b-8c99-4863-80de-41146440cf57-kube-api-access-twt4c\") pod \"nova-api-0\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.845416 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.924274 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:51:17 crc kubenswrapper[4912]: W1203 00:51:17.933059 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8105f90c_0397_4120_b23d_a0d3d72f6b01.slice/crio-a5dd41fead2fd7e1aac2078aaf3e87e8fb9f28be2c5e9ef43b7e9e92389142e8 WatchSource:0}: Error finding container a5dd41fead2fd7e1aac2078aaf3e87e8fb9f28be2c5e9ef43b7e9e92389142e8: Status 404 returned error can't find the container with id a5dd41fead2fd7e1aac2078aaf3e87e8fb9f28be2c5e9ef43b7e9e92389142e8 Dec 03 00:51:17 crc kubenswrapper[4912]: I1203 00:51:17.995385 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8105f90c-0397-4120-b23d-a0d3d72f6b01","Type":"ContainerStarted","Data":"a5dd41fead2fd7e1aac2078aaf3e87e8fb9f28be2c5e9ef43b7e9e92389142e8"} Dec 03 00:51:18 crc kubenswrapper[4912]: I1203 00:51:18.340293 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:18 crc kubenswrapper[4912]: I1203 00:51:18.589113 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35ff0444-86e0-4755-827a-2abbf5d42256" path="/var/lib/kubelet/pods/35ff0444-86e0-4755-827a-2abbf5d42256/volumes" Dec 03 00:51:18 crc kubenswrapper[4912]: I1203 00:51:18.591099 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70e7336c-9475-4488-bf3f-5768fece3a8f" path="/var/lib/kubelet/pods/70e7336c-9475-4488-bf3f-5768fece3a8f/volumes" Dec 03 00:51:19 crc kubenswrapper[4912]: I1203 00:51:19.010206 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80f6c2b-8c99-4863-80de-41146440cf57","Type":"ContainerStarted","Data":"7fc35fed2e03e7c1594b7ffbb3ae3e7aafcdd4709446d217834465a79e9a786f"} Dec 03 00:51:19 crc kubenswrapper[4912]: I1203 00:51:19.011473 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80f6c2b-8c99-4863-80de-41146440cf57","Type":"ContainerStarted","Data":"b28aaa53b1764392a92e21d44eda3abc76eab8452f72f75e2c17b0064cfe30d8"} Dec 03 00:51:19 crc kubenswrapper[4912]: I1203 00:51:19.011545 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80f6c2b-8c99-4863-80de-41146440cf57","Type":"ContainerStarted","Data":"9345d196d6f4f5b1e93b6ccc0f1f8657d4170c87c79cecbd073bd23dc3673455"} Dec 03 00:51:19 crc kubenswrapper[4912]: I1203 00:51:19.013488 4912 generic.go:334] "Generic (PLEG): container finished" podID="cfd2dfcd-2539-425b-9071-d0de6bc9a598" containerID="9b9e8f9697d2a56d7564f1fd10315324cd057bf556cf51049d076bf7644a6aa3" exitCode=0 Dec 03 00:51:19 crc kubenswrapper[4912]: I1203 00:51:19.013585 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-z6bqt" event={"ID":"cfd2dfcd-2539-425b-9071-d0de6bc9a598","Type":"ContainerDied","Data":"9b9e8f9697d2a56d7564f1fd10315324cd057bf556cf51049d076bf7644a6aa3"} Dec 03 00:51:19 crc kubenswrapper[4912]: I1203 00:51:19.015591 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8105f90c-0397-4120-b23d-a0d3d72f6b01","Type":"ContainerStarted","Data":"2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424"} Dec 03 00:51:19 crc kubenswrapper[4912]: I1203 00:51:19.053957 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.05392394 podStartE2EDuration="2.05392394s" podCreationTimestamp="2025-12-03 00:51:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:19.034517141 +0000 UTC m=+1664.676537761" watchObservedRunningTime="2025-12-03 00:51:19.05392394 +0000 UTC m=+1664.695944510" Dec 03 00:51:19 crc kubenswrapper[4912]: I1203 00:51:19.081853 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.081819455 podStartE2EDuration="2.081819455s" podCreationTimestamp="2025-12-03 00:51:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:19.065584171 +0000 UTC m=+1664.707604801" watchObservedRunningTime="2025-12-03 00:51:19.081819455 +0000 UTC m=+1664.723840055" Dec 03 00:51:19 crc kubenswrapper[4912]: I1203 00:51:19.836632 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.567104 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.606262 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-scripts\") pod \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.606449 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-config-data\") pod \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.606492 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r5rn\" (UniqueName: \"kubernetes.io/projected/cfd2dfcd-2539-425b-9071-d0de6bc9a598-kube-api-access-8r5rn\") pod \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.606595 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-combined-ca-bundle\") pod \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\" (UID: \"cfd2dfcd-2539-425b-9071-d0de6bc9a598\") " Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.654703 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-scripts" (OuterVolumeSpecName: "scripts") pod "cfd2dfcd-2539-425b-9071-d0de6bc9a598" (UID: "cfd2dfcd-2539-425b-9071-d0de6bc9a598"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.658900 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfd2dfcd-2539-425b-9071-d0de6bc9a598-kube-api-access-8r5rn" (OuterVolumeSpecName: "kube-api-access-8r5rn") pod "cfd2dfcd-2539-425b-9071-d0de6bc9a598" (UID: "cfd2dfcd-2539-425b-9071-d0de6bc9a598"). InnerVolumeSpecName "kube-api-access-8r5rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.664311 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-config-data" (OuterVolumeSpecName: "config-data") pod "cfd2dfcd-2539-425b-9071-d0de6bc9a598" (UID: "cfd2dfcd-2539-425b-9071-d0de6bc9a598"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.664402 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfd2dfcd-2539-425b-9071-d0de6bc9a598" (UID: "cfd2dfcd-2539-425b-9071-d0de6bc9a598"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.710062 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.710108 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.710127 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r5rn\" (UniqueName: \"kubernetes.io/projected/cfd2dfcd-2539-425b-9071-d0de6bc9a598-kube-api-access-8r5rn\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:20 crc kubenswrapper[4912]: I1203 00:51:20.710141 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfd2dfcd-2539-425b-9071-d0de6bc9a598-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:21 crc kubenswrapper[4912]: I1203 00:51:21.052569 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-z6bqt" event={"ID":"cfd2dfcd-2539-425b-9071-d0de6bc9a598","Type":"ContainerDied","Data":"0b1985743d6271b8708817b58d8e54b04f189a416b56a81379bc938e2420789e"} Dec 03 00:51:21 crc kubenswrapper[4912]: I1203 00:51:21.052661 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b1985743d6271b8708817b58d8e54b04f189a416b56a81379bc938e2420789e" Dec 03 00:51:21 crc kubenswrapper[4912]: I1203 00:51:21.052684 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-z6bqt" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.392298 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.425972 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.599933 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:51:22 crc kubenswrapper[4912]: E1203 00:51:22.600362 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.649846 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 00:51:22 crc kubenswrapper[4912]: E1203 00:51:22.650282 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfd2dfcd-2539-425b-9071-d0de6bc9a598" containerName="aodh-db-sync" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.650295 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfd2dfcd-2539-425b-9071-d0de6bc9a598" containerName="aodh-db-sync" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.650558 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfd2dfcd-2539-425b-9071-d0de6bc9a598" containerName="aodh-db-sync" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.653107 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.653214 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.656145 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tnsxp" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.656330 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.662216 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.667210 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-scripts\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.667270 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.667352 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq692\" (UniqueName: \"kubernetes.io/projected/920fd13c-46b2-4fad-a270-c9c900b913f1-kube-api-access-cq692\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.667393 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-config-data\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.770049 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-scripts\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.770136 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.770241 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq692\" (UniqueName: \"kubernetes.io/projected/920fd13c-46b2-4fad-a270-c9c900b913f1-kube-api-access-cq692\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.770291 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-config-data\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.777840 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.778066 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-config-data\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.779820 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-scripts\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.789256 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq692\" (UniqueName: \"kubernetes.io/projected/920fd13c-46b2-4fad-a270-c9c900b913f1-kube-api-access-cq692\") pod \"aodh-0\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " pod="openstack/aodh-0" Dec 03 00:51:22 crc kubenswrapper[4912]: I1203 00:51:22.996121 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:51:23 crc kubenswrapper[4912]: I1203 00:51:23.574260 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 00:51:24 crc kubenswrapper[4912]: I1203 00:51:24.107181 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerStarted","Data":"c2d3453912a8b7ca1609b55e6d50c64633c05409c8fe7ccaa0a1a9201d984784"} Dec 03 00:51:24 crc kubenswrapper[4912]: I1203 00:51:24.963843 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.032609 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-config-data\") pod \"863070c5-4b69-4066-af8f-109b2c9510fb\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.032789 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-scripts\") pod \"863070c5-4b69-4066-af8f-109b2c9510fb\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.032860 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-log-httpd\") pod \"863070c5-4b69-4066-af8f-109b2c9510fb\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.032898 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-run-httpd\") pod \"863070c5-4b69-4066-af8f-109b2c9510fb\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.032987 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2gfj\" (UniqueName: \"kubernetes.io/projected/863070c5-4b69-4066-af8f-109b2c9510fb-kube-api-access-l2gfj\") pod \"863070c5-4b69-4066-af8f-109b2c9510fb\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.033219 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-sg-core-conf-yaml\") pod \"863070c5-4b69-4066-af8f-109b2c9510fb\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.033237 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-combined-ca-bundle\") pod \"863070c5-4b69-4066-af8f-109b2c9510fb\" (UID: \"863070c5-4b69-4066-af8f-109b2c9510fb\") " Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.035552 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "863070c5-4b69-4066-af8f-109b2c9510fb" (UID: "863070c5-4b69-4066-af8f-109b2c9510fb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.037070 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "863070c5-4b69-4066-af8f-109b2c9510fb" (UID: "863070c5-4b69-4066-af8f-109b2c9510fb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.045109 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-scripts" (OuterVolumeSpecName: "scripts") pod "863070c5-4b69-4066-af8f-109b2c9510fb" (UID: "863070c5-4b69-4066-af8f-109b2c9510fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.054749 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/863070c5-4b69-4066-af8f-109b2c9510fb-kube-api-access-l2gfj" (OuterVolumeSpecName: "kube-api-access-l2gfj") pod "863070c5-4b69-4066-af8f-109b2c9510fb" (UID: "863070c5-4b69-4066-af8f-109b2c9510fb"). InnerVolumeSpecName "kube-api-access-l2gfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.099632 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "863070c5-4b69-4066-af8f-109b2c9510fb" (UID: "863070c5-4b69-4066-af8f-109b2c9510fb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.160857 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.161096 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/863070c5-4b69-4066-af8f-109b2c9510fb-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.161110 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2gfj\" (UniqueName: \"kubernetes.io/projected/863070c5-4b69-4066-af8f-109b2c9510fb-kube-api-access-l2gfj\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.161125 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.161134 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.174886 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerStarted","Data":"fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f"} Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.181146 4912 generic.go:334] "Generic (PLEG): container finished" podID="863070c5-4b69-4066-af8f-109b2c9510fb" containerID="2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c" exitCode=137 Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.181210 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerDied","Data":"2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c"} Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.181251 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"863070c5-4b69-4066-af8f-109b2c9510fb","Type":"ContainerDied","Data":"53d15e530a7b9b5c965a91aefae0348bb35a20b6ec171070f44a971477a5754e"} Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.181273 4912 scope.go:117] "RemoveContainer" containerID="2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.181582 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.228382 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "863070c5-4b69-4066-af8f-109b2c9510fb" (UID: "863070c5-4b69-4066-af8f-109b2c9510fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.262982 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.279669 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-config-data" (OuterVolumeSpecName: "config-data") pod "863070c5-4b69-4066-af8f-109b2c9510fb" (UID: "863070c5-4b69-4066-af8f-109b2c9510fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.305647 4912 scope.go:117] "RemoveContainer" containerID="7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.332522 4912 scope.go:117] "RemoveContainer" containerID="cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.365113 4912 scope.go:117] "RemoveContainer" containerID="9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.366960 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/863070c5-4b69-4066-af8f-109b2c9510fb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.394587 4912 scope.go:117] "RemoveContainer" containerID="2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c" Dec 03 00:51:25 crc kubenswrapper[4912]: E1203 00:51:25.395723 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c\": container with ID starting with 2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c not found: ID does not exist" containerID="2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.395823 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c"} err="failed to get container status \"2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c\": rpc error: code = NotFound desc = could not find container \"2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c\": container with ID starting with 2085933cb8a4e134f2135ff3289f3663b71563832bc01d9577c2c5009172101c not found: ID does not exist" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.395911 4912 scope.go:117] "RemoveContainer" containerID="7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab" Dec 03 00:51:25 crc kubenswrapper[4912]: E1203 00:51:25.396492 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab\": container with ID starting with 7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab not found: ID does not exist" containerID="7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.396522 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab"} err="failed to get container status \"7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab\": rpc error: code = NotFound desc = could not find container \"7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab\": container with ID starting with 7a31b4ed69b909d4c481d2471e8333a41920cf74a6eb6de67a5b8a9e86af31ab not found: ID does not exist" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.396539 4912 scope.go:117] "RemoveContainer" containerID="cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23" Dec 03 00:51:25 crc kubenswrapper[4912]: E1203 00:51:25.396850 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23\": container with ID starting with cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23 not found: ID does not exist" containerID="cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.396964 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23"} err="failed to get container status \"cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23\": rpc error: code = NotFound desc = could not find container \"cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23\": container with ID starting with cbf64f345a5025bc774e6abe580a9dd1570110b30ea6432c0a718904cbe83e23 not found: ID does not exist" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.397052 4912 scope.go:117] "RemoveContainer" containerID="9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf" Dec 03 00:51:25 crc kubenswrapper[4912]: E1203 00:51:25.397380 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf\": container with ID starting with 9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf not found: ID does not exist" containerID="9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.397412 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf"} err="failed to get container status \"9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf\": rpc error: code = NotFound desc = could not find container \"9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf\": container with ID starting with 9f7e5a1fac03693b9f1f343064ae1fd98ca93201ddc4c576670ee99a52fdaadf not found: ID does not exist" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.536213 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.557844 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.582444 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:25 crc kubenswrapper[4912]: E1203 00:51:25.583045 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="ceilometer-notification-agent" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.583068 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="ceilometer-notification-agent" Dec 03 00:51:25 crc kubenswrapper[4912]: E1203 00:51:25.583087 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="ceilometer-central-agent" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.583095 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="ceilometer-central-agent" Dec 03 00:51:25 crc kubenswrapper[4912]: E1203 00:51:25.583113 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="proxy-httpd" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.583119 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="proxy-httpd" Dec 03 00:51:25 crc kubenswrapper[4912]: E1203 00:51:25.583139 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="sg-core" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.583145 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="sg-core" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.583354 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="sg-core" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.583380 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="ceilometer-notification-agent" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.583396 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="proxy-httpd" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.583408 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" containerName="ceilometer-central-agent" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.585532 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.597752 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.598173 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.599027 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.682011 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psxjr\" (UniqueName: \"kubernetes.io/projected/86a75329-7796-40c9-a440-9ea7a2214221-kube-api-access-psxjr\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.682618 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-scripts\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.682805 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-config-data\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.683096 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-run-httpd\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.683171 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.683241 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.683549 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-log-httpd\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.785661 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-run-httpd\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.785752 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.785794 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.785887 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-log-httpd\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.785952 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psxjr\" (UniqueName: \"kubernetes.io/projected/86a75329-7796-40c9-a440-9ea7a2214221-kube-api-access-psxjr\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.785984 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-scripts\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.786054 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-config-data\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.786392 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-run-httpd\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.786879 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-log-httpd\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.800620 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.800752 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-scripts\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.801389 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.804087 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-config-data\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.807520 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psxjr\" (UniqueName: \"kubernetes.io/projected/86a75329-7796-40c9-a440-9ea7a2214221-kube-api-access-psxjr\") pod \"ceilometer-0\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " pod="openstack/ceilometer-0" Dec 03 00:51:25 crc kubenswrapper[4912]: I1203 00:51:25.917958 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:26 crc kubenswrapper[4912]: I1203 00:51:26.380306 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:26 crc kubenswrapper[4912]: I1203 00:51:26.600243 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="863070c5-4b69-4066-af8f-109b2c9510fb" path="/var/lib/kubelet/pods/863070c5-4b69-4066-af8f-109b2c9510fb/volumes" Dec 03 00:51:26 crc kubenswrapper[4912]: I1203 00:51:26.811345 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 00:51:26 crc kubenswrapper[4912]: W1203 00:51:26.911893 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86a75329_7796_40c9_a440_9ea7a2214221.slice/crio-a7b3a51c4329dea6acf34ec67dff412de68acb1ef17b095a727631fcc7ab5679 WatchSource:0}: Error finding container a7b3a51c4329dea6acf34ec67dff412de68acb1ef17b095a727631fcc7ab5679: Status 404 returned error can't find the container with id a7b3a51c4329dea6acf34ec67dff412de68acb1ef17b095a727631fcc7ab5679 Dec 03 00:51:26 crc kubenswrapper[4912]: I1203 00:51:26.914616 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:27 crc kubenswrapper[4912]: I1203 00:51:27.208270 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerStarted","Data":"b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1"} Dec 03 00:51:27 crc kubenswrapper[4912]: I1203 00:51:27.210464 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerStarted","Data":"a7b3a51c4329dea6acf34ec67dff412de68acb1ef17b095a727631fcc7ab5679"} Dec 03 00:51:27 crc kubenswrapper[4912]: I1203 00:51:27.395772 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 00:51:27 crc kubenswrapper[4912]: I1203 00:51:27.441154 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 00:51:27 crc kubenswrapper[4912]: I1203 00:51:27.846328 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 00:51:27 crc kubenswrapper[4912]: I1203 00:51:27.847052 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 00:51:28 crc kubenswrapper[4912]: I1203 00:51:28.233078 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerStarted","Data":"de24bdea9738f9522838f309e3c044192fb60cf529a4ed77f8d87e5149a86f3f"} Dec 03 00:51:28 crc kubenswrapper[4912]: I1203 00:51:28.280186 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 00:51:28 crc kubenswrapper[4912]: I1203 00:51:28.890567 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.228:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:51:28 crc kubenswrapper[4912]: I1203 00:51:28.890748 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.228:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:51:30 crc kubenswrapper[4912]: I1203 00:51:30.268416 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerStarted","Data":"03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127"} Dec 03 00:51:30 crc kubenswrapper[4912]: I1203 00:51:30.277819 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerStarted","Data":"1fda886ce68941b152cf2b3a169a898d7dfc219502309a3de8a678160b292723"} Dec 03 00:51:31 crc kubenswrapper[4912]: I1203 00:51:31.295087 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerStarted","Data":"ef5845a0059a090dced050d7e281002fabd7a1778d600194bcf496cee06bb74e"} Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.328748 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-api" containerID="cri-o://fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f" gracePeriod=30 Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.328869 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-notifier" containerID="cri-o://03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127" gracePeriod=30 Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.328965 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerStarted","Data":"54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e"} Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.328884 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-listener" containerID="cri-o://54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e" gracePeriod=30 Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.329008 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-evaluator" containerID="cri-o://b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1" gracePeriod=30 Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.342769 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerStarted","Data":"d22847b93bd539b0ad406d0e687bbde3b2ffb2c153b5889c5419764a02a5f584"} Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.343228 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="ceilometer-central-agent" containerID="cri-o://de24bdea9738f9522838f309e3c044192fb60cf529a4ed77f8d87e5149a86f3f" gracePeriod=30 Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.343405 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.343527 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="proxy-httpd" containerID="cri-o://d22847b93bd539b0ad406d0e687bbde3b2ffb2c153b5889c5419764a02a5f584" gracePeriod=30 Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.343620 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="sg-core" containerID="cri-o://ef5845a0059a090dced050d7e281002fabd7a1778d600194bcf496cee06bb74e" gracePeriod=30 Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.343774 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="ceilometer-notification-agent" containerID="cri-o://1fda886ce68941b152cf2b3a169a898d7dfc219502309a3de8a678160b292723" gracePeriod=30 Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.364901 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.742567951 podStartE2EDuration="11.364876443s" podCreationTimestamp="2025-12-03 00:51:22 +0000 UTC" firstStartedPulling="2025-12-03 00:51:23.531138116 +0000 UTC m=+1669.173158676" lastFinishedPulling="2025-12-03 00:51:32.153446608 +0000 UTC m=+1677.795467168" observedRunningTime="2025-12-03 00:51:33.356802865 +0000 UTC m=+1678.998823455" watchObservedRunningTime="2025-12-03 00:51:33.364876443 +0000 UTC m=+1679.006897003" Dec 03 00:51:33 crc kubenswrapper[4912]: I1203 00:51:33.407052 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.764981271 podStartE2EDuration="8.4070249s" podCreationTimestamp="2025-12-03 00:51:25 +0000 UTC" firstStartedPulling="2025-12-03 00:51:26.914600225 +0000 UTC m=+1672.556620785" lastFinishedPulling="2025-12-03 00:51:32.556643844 +0000 UTC m=+1678.198664414" observedRunningTime="2025-12-03 00:51:33.397048641 +0000 UTC m=+1679.039069211" watchObservedRunningTime="2025-12-03 00:51:33.4070249 +0000 UTC m=+1679.049045460" Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.361585 4912 generic.go:334] "Generic (PLEG): container finished" podID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerID="03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127" exitCode=0 Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.361634 4912 generic.go:334] "Generic (PLEG): container finished" podID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerID="b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1" exitCode=0 Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.361646 4912 generic.go:334] "Generic (PLEG): container finished" podID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerID="fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f" exitCode=0 Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.361660 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerDied","Data":"03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127"} Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.361710 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerDied","Data":"b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1"} Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.361725 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerDied","Data":"fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f"} Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.366296 4912 generic.go:334] "Generic (PLEG): container finished" podID="86a75329-7796-40c9-a440-9ea7a2214221" containerID="d22847b93bd539b0ad406d0e687bbde3b2ffb2c153b5889c5419764a02a5f584" exitCode=0 Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.366333 4912 generic.go:334] "Generic (PLEG): container finished" podID="86a75329-7796-40c9-a440-9ea7a2214221" containerID="ef5845a0059a090dced050d7e281002fabd7a1778d600194bcf496cee06bb74e" exitCode=2 Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.366346 4912 generic.go:334] "Generic (PLEG): container finished" podID="86a75329-7796-40c9-a440-9ea7a2214221" containerID="1fda886ce68941b152cf2b3a169a898d7dfc219502309a3de8a678160b292723" exitCode=0 Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.366353 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerDied","Data":"d22847b93bd539b0ad406d0e687bbde3b2ffb2c153b5889c5419764a02a5f584"} Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.366396 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerDied","Data":"ef5845a0059a090dced050d7e281002fabd7a1778d600194bcf496cee06bb74e"} Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.366412 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerDied","Data":"1fda886ce68941b152cf2b3a169a898d7dfc219502309a3de8a678160b292723"} Dec 03 00:51:34 crc kubenswrapper[4912]: I1203 00:51:34.591866 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:51:34 crc kubenswrapper[4912]: E1203 00:51:34.592637 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.323240 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.332304 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.391891 4912 generic.go:334] "Generic (PLEG): container finished" podID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerID="220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3" exitCode=137 Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.391969 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241c4c95-26fb-45ca-8a84-f414506c4d61","Type":"ContainerDied","Data":"220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3"} Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.392006 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"241c4c95-26fb-45ca-8a84-f414506c4d61","Type":"ContainerDied","Data":"df482892241c93c38533a50a5d50b24ed5680b5d7fb759c69d9a276636685f00"} Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.392024 4912 scope.go:117] "RemoveContainer" containerID="220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.392164 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.395707 4912 generic.go:334] "Generic (PLEG): container finished" podID="19d7c201-bd2f-4843-9a2f-eef0d9e547f9" containerID="d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67" exitCode=137 Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.395762 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"19d7c201-bd2f-4843-9a2f-eef0d9e547f9","Type":"ContainerDied","Data":"d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67"} Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.395801 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"19d7c201-bd2f-4843-9a2f-eef0d9e547f9","Type":"ContainerDied","Data":"83ccf98de4e84fbe28bfbe0aee9afcb9867e43be06d52112e79700a0996bd64d"} Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.395915 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.425629 4912 scope.go:117] "RemoveContainer" containerID="e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.452610 4912 scope.go:117] "RemoveContainer" containerID="220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3" Dec 03 00:51:35 crc kubenswrapper[4912]: E1203 00:51:35.453172 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3\": container with ID starting with 220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3 not found: ID does not exist" containerID="220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.453220 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3"} err="failed to get container status \"220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3\": rpc error: code = NotFound desc = could not find container \"220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3\": container with ID starting with 220d47fd2dd67e09834c96f521d2b2f161cb7d81bc04a60e5d1e78bc54db00a3 not found: ID does not exist" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.453249 4912 scope.go:117] "RemoveContainer" containerID="e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75" Dec 03 00:51:35 crc kubenswrapper[4912]: E1203 00:51:35.453821 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75\": container with ID starting with e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75 not found: ID does not exist" containerID="e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.453887 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75"} err="failed to get container status \"e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75\": rpc error: code = NotFound desc = could not find container \"e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75\": container with ID starting with e515edf49bd95bc3fe8d41fd56a2be681a2de22a0bcc8953f9556f62dad78a75 not found: ID does not exist" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.453932 4912 scope.go:117] "RemoveContainer" containerID="d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.454740 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-config-data\") pod \"241c4c95-26fb-45ca-8a84-f414506c4d61\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.454854 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-combined-ca-bundle\") pod \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.454903 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb5cb\" (UniqueName: \"kubernetes.io/projected/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-kube-api-access-pb5cb\") pod \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.454973 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-config-data\") pod \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\" (UID: \"19d7c201-bd2f-4843-9a2f-eef0d9e547f9\") " Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.455108 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slrbz\" (UniqueName: \"kubernetes.io/projected/241c4c95-26fb-45ca-8a84-f414506c4d61-kube-api-access-slrbz\") pod \"241c4c95-26fb-45ca-8a84-f414506c4d61\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.455247 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241c4c95-26fb-45ca-8a84-f414506c4d61-logs\") pod \"241c4c95-26fb-45ca-8a84-f414506c4d61\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.455333 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-combined-ca-bundle\") pod \"241c4c95-26fb-45ca-8a84-f414506c4d61\" (UID: \"241c4c95-26fb-45ca-8a84-f414506c4d61\") " Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.458032 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/241c4c95-26fb-45ca-8a84-f414506c4d61-logs" (OuterVolumeSpecName: "logs") pod "241c4c95-26fb-45ca-8a84-f414506c4d61" (UID: "241c4c95-26fb-45ca-8a84-f414506c4d61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.467783 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/241c4c95-26fb-45ca-8a84-f414506c4d61-kube-api-access-slrbz" (OuterVolumeSpecName: "kube-api-access-slrbz") pod "241c4c95-26fb-45ca-8a84-f414506c4d61" (UID: "241c4c95-26fb-45ca-8a84-f414506c4d61"). InnerVolumeSpecName "kube-api-access-slrbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.468330 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-kube-api-access-pb5cb" (OuterVolumeSpecName: "kube-api-access-pb5cb") pod "19d7c201-bd2f-4843-9a2f-eef0d9e547f9" (UID: "19d7c201-bd2f-4843-9a2f-eef0d9e547f9"). InnerVolumeSpecName "kube-api-access-pb5cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.504707 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-config-data" (OuterVolumeSpecName: "config-data") pod "241c4c95-26fb-45ca-8a84-f414506c4d61" (UID: "241c4c95-26fb-45ca-8a84-f414506c4d61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.510033 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "241c4c95-26fb-45ca-8a84-f414506c4d61" (UID: "241c4c95-26fb-45ca-8a84-f414506c4d61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.549630 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19d7c201-bd2f-4843-9a2f-eef0d9e547f9" (UID: "19d7c201-bd2f-4843-9a2f-eef0d9e547f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.554134 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-config-data" (OuterVolumeSpecName: "config-data") pod "19d7c201-bd2f-4843-9a2f-eef0d9e547f9" (UID: "19d7c201-bd2f-4843-9a2f-eef0d9e547f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.560015 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/241c4c95-26fb-45ca-8a84-f414506c4d61-logs\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.560061 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.560073 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/241c4c95-26fb-45ca-8a84-f414506c4d61-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.560082 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.560090 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb5cb\" (UniqueName: \"kubernetes.io/projected/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-kube-api-access-pb5cb\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.560099 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19d7c201-bd2f-4843-9a2f-eef0d9e547f9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.560112 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slrbz\" (UniqueName: \"kubernetes.io/projected/241c4c95-26fb-45ca-8a84-f414506c4d61-kube-api-access-slrbz\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.697598 4912 scope.go:117] "RemoveContainer" containerID="d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67" Dec 03 00:51:35 crc kubenswrapper[4912]: E1203 00:51:35.720925 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67\": container with ID starting with d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67 not found: ID does not exist" containerID="d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.720980 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67"} err="failed to get container status \"d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67\": rpc error: code = NotFound desc = could not find container \"d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67\": container with ID starting with d4a23ba81fd77a2aecf81235d91ba7164685e98422754b816fde7431b6c0ae67 not found: ID does not exist" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.785480 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.810307 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.829509 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.847505 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.881107 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:51:35 crc kubenswrapper[4912]: E1203 00:51:35.882331 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerName="nova-metadata-log" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.882363 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerName="nova-metadata-log" Dec 03 00:51:35 crc kubenswrapper[4912]: E1203 00:51:35.882446 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerName="nova-metadata-metadata" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.882468 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerName="nova-metadata-metadata" Dec 03 00:51:35 crc kubenswrapper[4912]: E1203 00:51:35.882530 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d7c201-bd2f-4843-9a2f-eef0d9e547f9" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.882545 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d7c201-bd2f-4843-9a2f-eef0d9e547f9" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.883155 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d7c201-bd2f-4843-9a2f-eef0d9e547f9" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.883193 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerName="nova-metadata-metadata" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.883215 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="241c4c95-26fb-45ca-8a84-f414506c4d61" containerName="nova-metadata-log" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.886176 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.891327 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.895013 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.912942 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.916249 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.923505 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.925026 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.923910 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.943525 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.955695 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.994767 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.994823 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vvkn\" (UniqueName: \"kubernetes.io/projected/ccac98c6-e328-46ea-b277-acf80bcc7dec-kube-api-access-7vvkn\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.994863 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccac98c6-e328-46ea-b277-acf80bcc7dec-logs\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.994894 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-config-data\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.994926 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.994974 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.994998 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqmrj\" (UniqueName: \"kubernetes.io/projected/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-kube-api-access-xqmrj\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.995028 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.995145 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:35 crc kubenswrapper[4912]: I1203 00:51:35.995166 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096032 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccac98c6-e328-46ea-b277-acf80bcc7dec-logs\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096095 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-config-data\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096139 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096197 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096217 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqmrj\" (UniqueName: \"kubernetes.io/projected/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-kube-api-access-xqmrj\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096246 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096338 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096363 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096410 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096453 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vvkn\" (UniqueName: \"kubernetes.io/projected/ccac98c6-e328-46ea-b277-acf80bcc7dec-kube-api-access-7vvkn\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.096522 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccac98c6-e328-46ea-b277-acf80bcc7dec-logs\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.113635 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.113673 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-config-data\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.113691 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.113795 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.114269 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.114290 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.114383 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.118701 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vvkn\" (UniqueName: \"kubernetes.io/projected/ccac98c6-e328-46ea-b277-acf80bcc7dec-kube-api-access-7vvkn\") pod \"nova-metadata-0\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.120640 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqmrj\" (UniqueName: \"kubernetes.io/projected/c73f5ab7-bc17-41fa-80af-55fbb2ebf0de-kube-api-access-xqmrj\") pod \"nova-cell1-novncproxy-0\" (UID: \"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.259632 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.271368 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.418602 4912 generic.go:334] "Generic (PLEG): container finished" podID="86a75329-7796-40c9-a440-9ea7a2214221" containerID="de24bdea9738f9522838f309e3c044192fb60cf529a4ed77f8d87e5149a86f3f" exitCode=0 Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.418711 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerDied","Data":"de24bdea9738f9522838f309e3c044192fb60cf529a4ed77f8d87e5149a86f3f"} Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.584419 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19d7c201-bd2f-4843-9a2f-eef0d9e547f9" path="/var/lib/kubelet/pods/19d7c201-bd2f-4843-9a2f-eef0d9e547f9/volumes" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.585889 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="241c4c95-26fb-45ca-8a84-f414506c4d61" path="/var/lib/kubelet/pods/241c4c95-26fb-45ca-8a84-f414506c4d61/volumes" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.931543 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:36 crc kubenswrapper[4912]: I1203 00:51:36.936665 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.023407 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-run-httpd\") pod \"86a75329-7796-40c9-a440-9ea7a2214221\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.023510 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-scripts\") pod \"86a75329-7796-40c9-a440-9ea7a2214221\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.023546 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-log-httpd\") pod \"86a75329-7796-40c9-a440-9ea7a2214221\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.023643 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-sg-core-conf-yaml\") pod \"86a75329-7796-40c9-a440-9ea7a2214221\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.023752 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psxjr\" (UniqueName: \"kubernetes.io/projected/86a75329-7796-40c9-a440-9ea7a2214221-kube-api-access-psxjr\") pod \"86a75329-7796-40c9-a440-9ea7a2214221\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.023834 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-config-data\") pod \"86a75329-7796-40c9-a440-9ea7a2214221\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.023884 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-combined-ca-bundle\") pod \"86a75329-7796-40c9-a440-9ea7a2214221\" (UID: \"86a75329-7796-40c9-a440-9ea7a2214221\") " Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.024035 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "86a75329-7796-40c9-a440-9ea7a2214221" (UID: "86a75329-7796-40c9-a440-9ea7a2214221"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.024160 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "86a75329-7796-40c9-a440-9ea7a2214221" (UID: "86a75329-7796-40c9-a440-9ea7a2214221"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.025207 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.025230 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/86a75329-7796-40c9-a440-9ea7a2214221-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.030661 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-scripts" (OuterVolumeSpecName: "scripts") pod "86a75329-7796-40c9-a440-9ea7a2214221" (UID: "86a75329-7796-40c9-a440-9ea7a2214221"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.038799 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86a75329-7796-40c9-a440-9ea7a2214221-kube-api-access-psxjr" (OuterVolumeSpecName: "kube-api-access-psxjr") pod "86a75329-7796-40c9-a440-9ea7a2214221" (UID: "86a75329-7796-40c9-a440-9ea7a2214221"). InnerVolumeSpecName "kube-api-access-psxjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.083938 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "86a75329-7796-40c9-a440-9ea7a2214221" (UID: "86a75329-7796-40c9-a440-9ea7a2214221"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.109782 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.127816 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.127857 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psxjr\" (UniqueName: \"kubernetes.io/projected/86a75329-7796-40c9-a440-9ea7a2214221-kube-api-access-psxjr\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.127869 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:37 crc kubenswrapper[4912]: W1203 00:51:37.127943 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccac98c6_e328_46ea_b277_acf80bcc7dec.slice/crio-2479014a749ed133f576bbe65359cc63c3e06e51dfff406da091079d3e2da61a WatchSource:0}: Error finding container 2479014a749ed133f576bbe65359cc63c3e06e51dfff406da091079d3e2da61a: Status 404 returned error can't find the container with id 2479014a749ed133f576bbe65359cc63c3e06e51dfff406da091079d3e2da61a Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.150732 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86a75329-7796-40c9-a440-9ea7a2214221" (UID: "86a75329-7796-40c9-a440-9ea7a2214221"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.192883 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-config-data" (OuterVolumeSpecName: "config-data") pod "86a75329-7796-40c9-a440-9ea7a2214221" (UID: "86a75329-7796-40c9-a440-9ea7a2214221"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.230080 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.230117 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86a75329-7796-40c9-a440-9ea7a2214221-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.436918 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ccac98c6-e328-46ea-b277-acf80bcc7dec","Type":"ContainerStarted","Data":"5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d"} Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.436975 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ccac98c6-e328-46ea-b277-acf80bcc7dec","Type":"ContainerStarted","Data":"2479014a749ed133f576bbe65359cc63c3e06e51dfff406da091079d3e2da61a"} Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.438552 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de","Type":"ContainerStarted","Data":"e339a5ad43a50e36a53cb6e6c6acc0c90391407ce5d56a0e513e008a3ed3f999"} Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.438609 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c73f5ab7-bc17-41fa-80af-55fbb2ebf0de","Type":"ContainerStarted","Data":"21d78f324a7650df7fd9c32e0184cb1dc2cbdcfe41c73f833c8a8b22e3042f56"} Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.446926 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"86a75329-7796-40c9-a440-9ea7a2214221","Type":"ContainerDied","Data":"a7b3a51c4329dea6acf34ec67dff412de68acb1ef17b095a727631fcc7ab5679"} Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.447003 4912 scope.go:117] "RemoveContainer" containerID="d22847b93bd539b0ad406d0e687bbde3b2ffb2c153b5889c5419764a02a5f584" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.447285 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.476023 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.475995658 podStartE2EDuration="2.475995658s" podCreationTimestamp="2025-12-03 00:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:37.461997641 +0000 UTC m=+1683.104018221" watchObservedRunningTime="2025-12-03 00:51:37.475995658 +0000 UTC m=+1683.118016218" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.630284 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.640110 4912 scope.go:117] "RemoveContainer" containerID="ef5845a0059a090dced050d7e281002fabd7a1778d600194bcf496cee06bb74e" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.676253 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.691709 4912 scope.go:117] "RemoveContainer" containerID="1fda886ce68941b152cf2b3a169a898d7dfc219502309a3de8a678160b292723" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.698310 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:37 crc kubenswrapper[4912]: E1203 00:51:37.698977 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="ceilometer-central-agent" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.699004 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="ceilometer-central-agent" Dec 03 00:51:37 crc kubenswrapper[4912]: E1203 00:51:37.699024 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="sg-core" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.699032 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="sg-core" Dec 03 00:51:37 crc kubenswrapper[4912]: E1203 00:51:37.699069 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="proxy-httpd" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.699075 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="proxy-httpd" Dec 03 00:51:37 crc kubenswrapper[4912]: E1203 00:51:37.699100 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="ceilometer-notification-agent" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.699107 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="ceilometer-notification-agent" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.699409 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="ceilometer-central-agent" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.699443 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="proxy-httpd" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.699459 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="ceilometer-notification-agent" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.699475 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="86a75329-7796-40c9-a440-9ea7a2214221" containerName="sg-core" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.706303 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.713918 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.714083 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.717377 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.742148 4912 scope.go:117] "RemoveContainer" containerID="de24bdea9738f9522838f309e3c044192fb60cf529a4ed77f8d87e5149a86f3f" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.746689 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wlf8\" (UniqueName: \"kubernetes.io/projected/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-kube-api-access-7wlf8\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.746879 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-scripts\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.747023 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-log-httpd\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.747096 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-config-data\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.747137 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.747349 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.747793 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-run-httpd\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.851303 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.851355 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.851407 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-run-httpd\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.851844 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wlf8\" (UniqueName: \"kubernetes.io/projected/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-kube-api-access-7wlf8\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.851962 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-scripts\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.852098 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-log-httpd\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.852148 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-config-data\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.852177 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.853724 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.861813 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.862304 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-run-httpd\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.862886 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.862959 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-log-httpd\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.863086 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.866701 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.868827 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-scripts\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.870635 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-config-data\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:37 crc kubenswrapper[4912]: I1203 00:51:37.887830 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wlf8\" (UniqueName: \"kubernetes.io/projected/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-kube-api-access-7wlf8\") pod \"ceilometer-0\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " pod="openstack/ceilometer-0" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.054021 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.467165 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ccac98c6-e328-46ea-b277-acf80bcc7dec","Type":"ContainerStarted","Data":"b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3"} Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.471615 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.482168 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.492971 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.492950807 podStartE2EDuration="3.492950807s" podCreationTimestamp="2025-12-03 00:51:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:38.489012142 +0000 UTC m=+1684.131032712" watchObservedRunningTime="2025-12-03 00:51:38.492950807 +0000 UTC m=+1684.134971367" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.585877 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86a75329-7796-40c9-a440-9ea7a2214221" path="/var/lib/kubelet/pods/86a75329-7796-40c9-a440-9ea7a2214221/volumes" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.663881 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.710646 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-kdbv2"] Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.719064 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.736488 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-kdbv2"] Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.784242 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx9m2\" (UniqueName: \"kubernetes.io/projected/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-kube-api-access-gx9m2\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.784343 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.784373 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.784414 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.784450 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.784501 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-config\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.886964 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.887030 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.887078 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.887107 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.887163 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-config\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.887244 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx9m2\" (UniqueName: \"kubernetes.io/projected/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-kube-api-access-gx9m2\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.888275 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.888650 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.888936 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.890091 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.891414 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-config\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:38 crc kubenswrapper[4912]: I1203 00:51:38.920932 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx9m2\" (UniqueName: \"kubernetes.io/projected/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-kube-api-access-gx9m2\") pod \"dnsmasq-dns-f84f9ccf-kdbv2\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:39 crc kubenswrapper[4912]: I1203 00:51:39.057680 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:39 crc kubenswrapper[4912]: I1203 00:51:39.558373 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerStarted","Data":"104368ece8698cc1226788dac133abef60911db22e0e8c84dea4b1eb667e3b46"} Dec 03 00:51:40 crc kubenswrapper[4912]: I1203 00:51:39.897883 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-kdbv2"] Dec 03 00:51:40 crc kubenswrapper[4912]: I1203 00:51:40.588684 4912 generic.go:334] "Generic (PLEG): container finished" podID="e700236d-b3e1-40b7-8e95-9e2d799bf0d9" containerID="dacaa5226c583e45648c8127972fbbf6adbb1b2e6dcf39b886cd09669f3242aa" exitCode=0 Dec 03 00:51:40 crc kubenswrapper[4912]: I1203 00:51:40.596704 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerStarted","Data":"74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564"} Dec 03 00:51:40 crc kubenswrapper[4912]: I1203 00:51:40.596764 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerStarted","Data":"6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687"} Dec 03 00:51:40 crc kubenswrapper[4912]: I1203 00:51:40.596780 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" event={"ID":"e700236d-b3e1-40b7-8e95-9e2d799bf0d9","Type":"ContainerDied","Data":"dacaa5226c583e45648c8127972fbbf6adbb1b2e6dcf39b886cd09669f3242aa"} Dec 03 00:51:40 crc kubenswrapper[4912]: I1203 00:51:40.596796 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" event={"ID":"e700236d-b3e1-40b7-8e95-9e2d799bf0d9","Type":"ContainerStarted","Data":"77689f23753c1140ac6f1fd94eb51cd6ec2c7bf8bbf2f3abe8666be32cc34e1b"} Dec 03 00:51:41 crc kubenswrapper[4912]: I1203 00:51:41.259898 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 00:51:41 crc kubenswrapper[4912]: I1203 00:51:41.260464 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 00:51:41 crc kubenswrapper[4912]: I1203 00:51:41.272634 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:41 crc kubenswrapper[4912]: I1203 00:51:41.603641 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" event={"ID":"e700236d-b3e1-40b7-8e95-9e2d799bf0d9","Type":"ContainerStarted","Data":"3c603903509e566fa9464d11c926c221ae776e6f127fc32229f3c6b5e8e535b6"} Dec 03 00:51:41 crc kubenswrapper[4912]: I1203 00:51:41.603832 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:41 crc kubenswrapper[4912]: I1203 00:51:41.608120 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerStarted","Data":"2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f"} Dec 03 00:51:41 crc kubenswrapper[4912]: I1203 00:51:41.640729 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" podStartSLOduration=3.6407071699999998 podStartE2EDuration="3.64070717s" podCreationTimestamp="2025-12-03 00:51:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:41.626767334 +0000 UTC m=+1687.268787914" watchObservedRunningTime="2025-12-03 00:51:41.64070717 +0000 UTC m=+1687.282727730" Dec 03 00:51:42 crc kubenswrapper[4912]: I1203 00:51:42.340629 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:42 crc kubenswrapper[4912]: I1203 00:51:42.341290 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-log" containerID="cri-o://b28aaa53b1764392a92e21d44eda3abc76eab8452f72f75e2c17b0064cfe30d8" gracePeriod=30 Dec 03 00:51:42 crc kubenswrapper[4912]: I1203 00:51:42.341369 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-api" containerID="cri-o://7fc35fed2e03e7c1594b7ffbb3ae3e7aafcdd4709446d217834465a79e9a786f" gracePeriod=30 Dec 03 00:51:42 crc kubenswrapper[4912]: I1203 00:51:42.630324 4912 generic.go:334] "Generic (PLEG): container finished" podID="b80f6c2b-8c99-4863-80de-41146440cf57" containerID="b28aaa53b1764392a92e21d44eda3abc76eab8452f72f75e2c17b0064cfe30d8" exitCode=143 Dec 03 00:51:42 crc kubenswrapper[4912]: I1203 00:51:42.631550 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80f6c2b-8c99-4863-80de-41146440cf57","Type":"ContainerDied","Data":"b28aaa53b1764392a92e21d44eda3abc76eab8452f72f75e2c17b0064cfe30d8"} Dec 03 00:51:43 crc kubenswrapper[4912]: I1203 00:51:43.645846 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerStarted","Data":"c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47"} Dec 03 00:51:43 crc kubenswrapper[4912]: I1203 00:51:43.646212 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:51:43 crc kubenswrapper[4912]: I1203 00:51:43.680168 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.773693792 podStartE2EDuration="6.680143827s" podCreationTimestamp="2025-12-03 00:51:37 +0000 UTC" firstStartedPulling="2025-12-03 00:51:38.692035968 +0000 UTC m=+1684.334056528" lastFinishedPulling="2025-12-03 00:51:42.598486003 +0000 UTC m=+1688.240506563" observedRunningTime="2025-12-03 00:51:43.670869687 +0000 UTC m=+1689.312890257" watchObservedRunningTime="2025-12-03 00:51:43.680143827 +0000 UTC m=+1689.322164387" Dec 03 00:51:44 crc kubenswrapper[4912]: I1203 00:51:44.210943 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:45 crc kubenswrapper[4912]: I1203 00:51:45.708872 4912 generic.go:334] "Generic (PLEG): container finished" podID="b80f6c2b-8c99-4863-80de-41146440cf57" containerID="7fc35fed2e03e7c1594b7ffbb3ae3e7aafcdd4709446d217834465a79e9a786f" exitCode=0 Dec 03 00:51:45 crc kubenswrapper[4912]: I1203 00:51:45.708956 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80f6c2b-8c99-4863-80de-41146440cf57","Type":"ContainerDied","Data":"7fc35fed2e03e7c1594b7ffbb3ae3e7aafcdd4709446d217834465a79e9a786f"} Dec 03 00:51:45 crc kubenswrapper[4912]: I1203 00:51:45.709715 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="proxy-httpd" containerID="cri-o://c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47" gracePeriod=30 Dec 03 00:51:45 crc kubenswrapper[4912]: I1203 00:51:45.709749 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="sg-core" containerID="cri-o://2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f" gracePeriod=30 Dec 03 00:51:45 crc kubenswrapper[4912]: I1203 00:51:45.709749 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="ceilometer-notification-agent" containerID="cri-o://74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564" gracePeriod=30 Dec 03 00:51:45 crc kubenswrapper[4912]: I1203 00:51:45.709887 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="ceilometer-central-agent" containerID="cri-o://6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687" gracePeriod=30 Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.235315 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.260985 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.262838 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.272394 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.297795 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-combined-ca-bundle\") pod \"b80f6c2b-8c99-4863-80de-41146440cf57\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.297964 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80f6c2b-8c99-4863-80de-41146440cf57-logs\") pod \"b80f6c2b-8c99-4863-80de-41146440cf57\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.298123 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-config-data\") pod \"b80f6c2b-8c99-4863-80de-41146440cf57\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.298210 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twt4c\" (UniqueName: \"kubernetes.io/projected/b80f6c2b-8c99-4863-80de-41146440cf57-kube-api-access-twt4c\") pod \"b80f6c2b-8c99-4863-80de-41146440cf57\" (UID: \"b80f6c2b-8c99-4863-80de-41146440cf57\") " Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.300172 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b80f6c2b-8c99-4863-80de-41146440cf57-logs" (OuterVolumeSpecName: "logs") pod "b80f6c2b-8c99-4863-80de-41146440cf57" (UID: "b80f6c2b-8c99-4863-80de-41146440cf57"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.307256 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b80f6c2b-8c99-4863-80de-41146440cf57-kube-api-access-twt4c" (OuterVolumeSpecName: "kube-api-access-twt4c") pod "b80f6c2b-8c99-4863-80de-41146440cf57" (UID: "b80f6c2b-8c99-4863-80de-41146440cf57"). InnerVolumeSpecName "kube-api-access-twt4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.323689 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.338881 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b80f6c2b-8c99-4863-80de-41146440cf57" (UID: "b80f6c2b-8c99-4863-80de-41146440cf57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.350230 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-config-data" (OuterVolumeSpecName: "config-data") pod "b80f6c2b-8c99-4863-80de-41146440cf57" (UID: "b80f6c2b-8c99-4863-80de-41146440cf57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.400908 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b80f6c2b-8c99-4863-80de-41146440cf57-logs\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.400956 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.400967 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twt4c\" (UniqueName: \"kubernetes.io/projected/b80f6c2b-8c99-4863-80de-41146440cf57-kube-api-access-twt4c\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.400976 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b80f6c2b-8c99-4863-80de-41146440cf57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.727011 4912 generic.go:334] "Generic (PLEG): container finished" podID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerID="c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47" exitCode=0 Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.727421 4912 generic.go:334] "Generic (PLEG): container finished" podID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerID="2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f" exitCode=2 Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.727450 4912 generic.go:334] "Generic (PLEG): container finished" podID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerID="74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564" exitCode=0 Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.727088 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerDied","Data":"c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47"} Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.727543 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerDied","Data":"2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f"} Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.727563 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerDied","Data":"74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564"} Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.732636 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.733284 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b80f6c2b-8c99-4863-80de-41146440cf57","Type":"ContainerDied","Data":"9345d196d6f4f5b1e93b6ccc0f1f8657d4170c87c79cecbd073bd23dc3673455"} Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.733328 4912 scope.go:117] "RemoveContainer" containerID="7fc35fed2e03e7c1594b7ffbb3ae3e7aafcdd4709446d217834465a79e9a786f" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.760537 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.765357 4912 scope.go:117] "RemoveContainer" containerID="b28aaa53b1764392a92e21d44eda3abc76eab8452f72f75e2c17b0064cfe30d8" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.766247 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.779936 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.830829 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:46 crc kubenswrapper[4912]: E1203 00:51:46.831530 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-api" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.831557 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-api" Dec 03 00:51:46 crc kubenswrapper[4912]: E1203 00:51:46.831577 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-log" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.831585 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-log" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.831889 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-api" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.831921 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" containerName="nova-api-log" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.843373 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.873777 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.873783 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.875018 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.891041 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.941620 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de4b39f9-20a8-4205-8131-50877d60d1f5-logs\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.942063 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b74mv\" (UniqueName: \"kubernetes.io/projected/de4b39f9-20a8-4205-8131-50877d60d1f5-kube-api-access-b74mv\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.942127 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.942930 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.943006 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-config-data\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:46 crc kubenswrapper[4912]: I1203 00:51:46.943026 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-public-tls-certs\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.045375 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-config-data\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.045456 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-public-tls-certs\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.045512 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de4b39f9-20a8-4205-8131-50877d60d1f5-logs\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.045676 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b74mv\" (UniqueName: \"kubernetes.io/projected/de4b39f9-20a8-4205-8131-50877d60d1f5-kube-api-access-b74mv\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.045739 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.045834 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.046393 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de4b39f9-20a8-4205-8131-50877d60d1f5-logs\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.052003 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-sgfjt"] Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.052995 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.053325 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-config-data\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.054027 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.054469 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.063052 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.063574 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.063929 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-public-tls-certs\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.077220 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-sgfjt"] Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.081056 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b74mv\" (UniqueName: \"kubernetes.io/projected/de4b39f9-20a8-4205-8131-50877d60d1f5-kube-api-access-b74mv\") pod \"nova-api-0\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.147810 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.147865 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmk9j\" (UniqueName: \"kubernetes.io/projected/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-kube-api-access-zmk9j\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.148354 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-config-data\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.148524 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-scripts\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.214728 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.251507 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-config-data\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.251622 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-scripts\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.251753 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.251786 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmk9j\" (UniqueName: \"kubernetes.io/projected/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-kube-api-access-zmk9j\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.263347 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-scripts\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.264209 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-config-data\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.267219 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.291557 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.231:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.291940 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.231:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.299273 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmk9j\" (UniqueName: \"kubernetes.io/projected/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-kube-api-access-zmk9j\") pod \"nova-cell1-cell-mapping-sgfjt\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.360757 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.581594 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:51:47 crc kubenswrapper[4912]: E1203 00:51:47.581991 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:51:47 crc kubenswrapper[4912]: I1203 00:51:47.981577 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:48 crc kubenswrapper[4912]: I1203 00:51:48.237504 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-sgfjt"] Dec 03 00:51:48 crc kubenswrapper[4912]: I1203 00:51:48.600638 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b80f6c2b-8c99-4863-80de-41146440cf57" path="/var/lib/kubelet/pods/b80f6c2b-8c99-4863-80de-41146440cf57/volumes" Dec 03 00:51:48 crc kubenswrapper[4912]: I1203 00:51:48.798732 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sgfjt" event={"ID":"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9","Type":"ContainerStarted","Data":"c92792f6d2e7d39a1ce700c1e84f032ce8f7da2035aeca94f08b7a6de4a897ca"} Dec 03 00:51:48 crc kubenswrapper[4912]: I1203 00:51:48.799120 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sgfjt" event={"ID":"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9","Type":"ContainerStarted","Data":"f817f9b3435a580a8b72e26f0ad68f863e9df7031f88520a40b1000c40137d54"} Dec 03 00:51:48 crc kubenswrapper[4912]: I1203 00:51:48.801988 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de4b39f9-20a8-4205-8131-50877d60d1f5","Type":"ContainerStarted","Data":"dd03310ff4d91964a6172396d8d73845fdb0fad9cced0ecfd0f286800339f81e"} Dec 03 00:51:48 crc kubenswrapper[4912]: I1203 00:51:48.802043 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de4b39f9-20a8-4205-8131-50877d60d1f5","Type":"ContainerStarted","Data":"3eab385cfdcd041284bf2e2e409dcaff795e0cd10d1dd1e84413a7b58443bef5"} Dec 03 00:51:48 crc kubenswrapper[4912]: I1203 00:51:48.802058 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de4b39f9-20a8-4205-8131-50877d60d1f5","Type":"ContainerStarted","Data":"1d5acafe73aed7932fc82fab964bdc93f99d4f546ee638ab14db2d2fcb52664d"} Dec 03 00:51:48 crc kubenswrapper[4912]: I1203 00:51:48.825956 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-sgfjt" podStartSLOduration=1.825928139 podStartE2EDuration="1.825928139s" podCreationTimestamp="2025-12-03 00:51:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:48.812920749 +0000 UTC m=+1694.454941309" watchObservedRunningTime="2025-12-03 00:51:48.825928139 +0000 UTC m=+1694.467948699" Dec 03 00:51:48 crc kubenswrapper[4912]: I1203 00:51:48.850980 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.850955754 podStartE2EDuration="2.850955754s" podCreationTimestamp="2025-12-03 00:51:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:51:48.836859304 +0000 UTC m=+1694.478879884" watchObservedRunningTime="2025-12-03 00:51:48.850955754 +0000 UTC m=+1694.492976314" Dec 03 00:51:49 crc kubenswrapper[4912]: I1203 00:51:49.060140 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:51:49 crc kubenswrapper[4912]: I1203 00:51:49.146408 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-pkxw5"] Dec 03 00:51:49 crc kubenswrapper[4912]: I1203 00:51:49.146718 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" podUID="e08a827e-2f84-4631-9c79-10c82eba397f" containerName="dnsmasq-dns" containerID="cri-o://4d8ef5d001aa2e826b865b2c60bc27c58ada9507f12132f98c5d6bbf394f8042" gracePeriod=10 Dec 03 00:51:49 crc kubenswrapper[4912]: I1203 00:51:49.822081 4912 generic.go:334] "Generic (PLEG): container finished" podID="e08a827e-2f84-4631-9c79-10c82eba397f" containerID="4d8ef5d001aa2e826b865b2c60bc27c58ada9507f12132f98c5d6bbf394f8042" exitCode=0 Dec 03 00:51:49 crc kubenswrapper[4912]: I1203 00:51:49.822247 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" event={"ID":"e08a827e-2f84-4631-9c79-10c82eba397f","Type":"ContainerDied","Data":"4d8ef5d001aa2e826b865b2c60bc27c58ada9507f12132f98c5d6bbf394f8042"} Dec 03 00:51:49 crc kubenswrapper[4912]: I1203 00:51:49.823471 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" event={"ID":"e08a827e-2f84-4631-9c79-10c82eba397f","Type":"ContainerDied","Data":"3c2c72f764fc32cb693994726a90e40ba8e00b3a8a3a58c1e25d91494668ccf8"} Dec 03 00:51:49 crc kubenswrapper[4912]: I1203 00:51:49.823485 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c2c72f764fc32cb693994726a90e40ba8e00b3a8a3a58c1e25d91494668ccf8" Dec 03 00:51:49 crc kubenswrapper[4912]: I1203 00:51:49.888584 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.081384 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt28k\" (UniqueName: \"kubernetes.io/projected/e08a827e-2f84-4631-9c79-10c82eba397f-kube-api-access-lt28k\") pod \"e08a827e-2f84-4631-9c79-10c82eba397f\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.081534 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-nb\") pod \"e08a827e-2f84-4631-9c79-10c82eba397f\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.081583 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-swift-storage-0\") pod \"e08a827e-2f84-4631-9c79-10c82eba397f\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.081631 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-sb\") pod \"e08a827e-2f84-4631-9c79-10c82eba397f\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.081743 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-config\") pod \"e08a827e-2f84-4631-9c79-10c82eba397f\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.081761 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-svc\") pod \"e08a827e-2f84-4631-9c79-10c82eba397f\" (UID: \"e08a827e-2f84-4631-9c79-10c82eba397f\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.097368 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e08a827e-2f84-4631-9c79-10c82eba397f-kube-api-access-lt28k" (OuterVolumeSpecName: "kube-api-access-lt28k") pod "e08a827e-2f84-4631-9c79-10c82eba397f" (UID: "e08a827e-2f84-4631-9c79-10c82eba397f"). InnerVolumeSpecName "kube-api-access-lt28k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.170070 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e08a827e-2f84-4631-9c79-10c82eba397f" (UID: "e08a827e-2f84-4631-9c79-10c82eba397f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.170919 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e08a827e-2f84-4631-9c79-10c82eba397f" (UID: "e08a827e-2f84-4631-9c79-10c82eba397f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.170965 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-config" (OuterVolumeSpecName: "config") pod "e08a827e-2f84-4631-9c79-10c82eba397f" (UID: "e08a827e-2f84-4631-9c79-10c82eba397f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.186331 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e08a827e-2f84-4631-9c79-10c82eba397f" (UID: "e08a827e-2f84-4631-9c79-10c82eba397f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.189803 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.189855 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.189872 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.189882 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.189894 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt28k\" (UniqueName: \"kubernetes.io/projected/e08a827e-2f84-4631-9c79-10c82eba397f-kube-api-access-lt28k\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.208061 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e08a827e-2f84-4631-9c79-10c82eba397f" (UID: "e08a827e-2f84-4631-9c79-10c82eba397f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.291928 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e08a827e-2f84-4631-9c79-10c82eba397f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.767159 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.873407 4912 generic.go:334] "Generic (PLEG): container finished" podID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerID="6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687" exitCode=0 Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.873554 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-pkxw5" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.874643 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.874888 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerDied","Data":"6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687"} Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.874925 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c61ad040-bde7-4aec-bb87-6cacd47ccd5a","Type":"ContainerDied","Data":"104368ece8698cc1226788dac133abef60911db22e0e8c84dea4b1eb667e3b46"} Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.874943 4912 scope.go:117] "RemoveContainer" containerID="c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.915415 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-scripts\") pod \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.915537 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-combined-ca-bundle\") pod \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.915616 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-log-httpd\") pod \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.915709 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-run-httpd\") pod \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.915781 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-config-data\") pod \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.915897 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-sg-core-conf-yaml\") pod \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.915934 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wlf8\" (UniqueName: \"kubernetes.io/projected/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-kube-api-access-7wlf8\") pod \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\" (UID: \"c61ad040-bde7-4aec-bb87-6cacd47ccd5a\") " Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.917782 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c61ad040-bde7-4aec-bb87-6cacd47ccd5a" (UID: "c61ad040-bde7-4aec-bb87-6cacd47ccd5a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.919403 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c61ad040-bde7-4aec-bb87-6cacd47ccd5a" (UID: "c61ad040-bde7-4aec-bb87-6cacd47ccd5a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.924306 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-scripts" (OuterVolumeSpecName: "scripts") pod "c61ad040-bde7-4aec-bb87-6cacd47ccd5a" (UID: "c61ad040-bde7-4aec-bb87-6cacd47ccd5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.927944 4912 scope.go:117] "RemoveContainer" containerID="2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.936174 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-pkxw5"] Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.952697 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-kube-api-access-7wlf8" (OuterVolumeSpecName: "kube-api-access-7wlf8") pod "c61ad040-bde7-4aec-bb87-6cacd47ccd5a" (UID: "c61ad040-bde7-4aec-bb87-6cacd47ccd5a"). InnerVolumeSpecName "kube-api-access-7wlf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.958666 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-pkxw5"] Dec 03 00:51:50 crc kubenswrapper[4912]: I1203 00:51:50.962291 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c61ad040-bde7-4aec-bb87-6cacd47ccd5a" (UID: "c61ad040-bde7-4aec-bb87-6cacd47ccd5a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.018833 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.018872 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wlf8\" (UniqueName: \"kubernetes.io/projected/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-kube-api-access-7wlf8\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.018883 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.018892 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.018903 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.063019 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c61ad040-bde7-4aec-bb87-6cacd47ccd5a" (UID: "c61ad040-bde7-4aec-bb87-6cacd47ccd5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.121627 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.156349 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-config-data" (OuterVolumeSpecName: "config-data") pod "c61ad040-bde7-4aec-bb87-6cacd47ccd5a" (UID: "c61ad040-bde7-4aec-bb87-6cacd47ccd5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.157782 4912 scope.go:117] "RemoveContainer" containerID="74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.183950 4912 scope.go:117] "RemoveContainer" containerID="6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.223018 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c61ad040-bde7-4aec-bb87-6cacd47ccd5a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.224680 4912 scope.go:117] "RemoveContainer" containerID="c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47" Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.227638 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47\": container with ID starting with c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47 not found: ID does not exist" containerID="c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.227706 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47"} err="failed to get container status \"c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47\": rpc error: code = NotFound desc = could not find container \"c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47\": container with ID starting with c8ca26d73730498ee432e9999d3fd725febfc5e9e544368b47b05ae3f6c82e47 not found: ID does not exist" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.227740 4912 scope.go:117] "RemoveContainer" containerID="2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f" Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.231592 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f\": container with ID starting with 2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f not found: ID does not exist" containerID="2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.231651 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f"} err="failed to get container status \"2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f\": rpc error: code = NotFound desc = could not find container \"2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f\": container with ID starting with 2fc9de33c8fa7c3abb1d183b92d99c35dc030c187608f49671d0090aa941fd5f not found: ID does not exist" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.231691 4912 scope.go:117] "RemoveContainer" containerID="74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564" Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.232398 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564\": container with ID starting with 74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564 not found: ID does not exist" containerID="74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.232462 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564"} err="failed to get container status \"74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564\": rpc error: code = NotFound desc = could not find container \"74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564\": container with ID starting with 74260461d7d9c8f3e120be8257407f1b957fefc7e66879e329337a628bdb4564 not found: ID does not exist" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.232498 4912 scope.go:117] "RemoveContainer" containerID="6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687" Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.232761 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687\": container with ID starting with 6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687 not found: ID does not exist" containerID="6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.232814 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687"} err="failed to get container status \"6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687\": rpc error: code = NotFound desc = could not find container \"6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687\": container with ID starting with 6f97ef520f434f8841bf062a3baf99a23c149b8cfd16d6c03fbade5d8cc11687 not found: ID does not exist" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.244359 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.260819 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.279366 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.279981 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="proxy-httpd" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.279998 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="proxy-httpd" Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.280020 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08a827e-2f84-4631-9c79-10c82eba397f" containerName="dnsmasq-dns" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.280027 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08a827e-2f84-4631-9c79-10c82eba397f" containerName="dnsmasq-dns" Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.280053 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="ceilometer-notification-agent" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.280061 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="ceilometer-notification-agent" Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.280081 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="sg-core" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.280087 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="sg-core" Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.280096 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="ceilometer-central-agent" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.280102 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="ceilometer-central-agent" Dec 03 00:51:51 crc kubenswrapper[4912]: E1203 00:51:51.280113 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e08a827e-2f84-4631-9c79-10c82eba397f" containerName="init" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.280119 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e08a827e-2f84-4631-9c79-10c82eba397f" containerName="init" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.283250 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="ceilometer-central-agent" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.283299 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="proxy-httpd" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.283310 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="ceilometer-notification-agent" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.283327 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" containerName="sg-core" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.283339 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="e08a827e-2f84-4631-9c79-10c82eba397f" containerName="dnsmasq-dns" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.285777 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.293027 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.294461 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.303495 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.428071 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-log-httpd\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.428138 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-run-httpd\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.428591 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-scripts\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.428708 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-config-data\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.428790 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6skmw\" (UniqueName: \"kubernetes.io/projected/a7a044e0-a3f5-462d-b348-20f1382d24dd-kube-api-access-6skmw\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.428952 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.428988 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.531760 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-scripts\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.531830 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-config-data\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.532060 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6skmw\" (UniqueName: \"kubernetes.io/projected/a7a044e0-a3f5-462d-b348-20f1382d24dd-kube-api-access-6skmw\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.532154 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.532182 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.532257 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-log-httpd\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.532297 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-run-httpd\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.532954 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-run-httpd\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.533386 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-log-httpd\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.536356 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.539409 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.540255 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-config-data\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.540940 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-scripts\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.556365 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6skmw\" (UniqueName: \"kubernetes.io/projected/a7a044e0-a3f5-462d-b348-20f1382d24dd-kube-api-access-6skmw\") pod \"ceilometer-0\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " pod="openstack/ceilometer-0" Dec 03 00:51:51 crc kubenswrapper[4912]: I1203 00:51:51.618445 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:51:52 crc kubenswrapper[4912]: W1203 00:51:52.201181 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7a044e0_a3f5_462d_b348_20f1382d24dd.slice/crio-b4c4ee09e7f88aa73f8cd72870bda2a5fc46656364fc3086fd45d706beac1f9b WatchSource:0}: Error finding container b4c4ee09e7f88aa73f8cd72870bda2a5fc46656364fc3086fd45d706beac1f9b: Status 404 returned error can't find the container with id b4c4ee09e7f88aa73f8cd72870bda2a5fc46656364fc3086fd45d706beac1f9b Dec 03 00:51:52 crc kubenswrapper[4912]: I1203 00:51:52.216906 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:51:52 crc kubenswrapper[4912]: I1203 00:51:52.592043 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c61ad040-bde7-4aec-bb87-6cacd47ccd5a" path="/var/lib/kubelet/pods/c61ad040-bde7-4aec-bb87-6cacd47ccd5a/volumes" Dec 03 00:51:52 crc kubenswrapper[4912]: I1203 00:51:52.593317 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e08a827e-2f84-4631-9c79-10c82eba397f" path="/var/lib/kubelet/pods/e08a827e-2f84-4631-9c79-10c82eba397f/volumes" Dec 03 00:51:52 crc kubenswrapper[4912]: I1203 00:51:52.907284 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerStarted","Data":"f37336a7f9a33dd358b2b1672a23e94279c67b8437827f79c8726ac95070ad3e"} Dec 03 00:51:52 crc kubenswrapper[4912]: I1203 00:51:52.907344 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerStarted","Data":"b4c4ee09e7f88aa73f8cd72870bda2a5fc46656364fc3086fd45d706beac1f9b"} Dec 03 00:51:53 crc kubenswrapper[4912]: I1203 00:51:53.922515 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerStarted","Data":"ae28558032828c1ff946132550939e18bcbf094f28fc719fc67fb5d85150bd99"} Dec 03 00:51:54 crc kubenswrapper[4912]: I1203 00:51:54.942874 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerStarted","Data":"5bed58584b46e63c99f836c9f9e2b3506135b293d4f3fa1a58851c90997349ba"} Dec 03 00:51:55 crc kubenswrapper[4912]: I1203 00:51:55.958542 4912 generic.go:334] "Generic (PLEG): container finished" podID="a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9" containerID="c92792f6d2e7d39a1ce700c1e84f032ce8f7da2035aeca94f08b7a6de4a897ca" exitCode=0 Dec 03 00:51:55 crc kubenswrapper[4912]: I1203 00:51:55.958620 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sgfjt" event={"ID":"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9","Type":"ContainerDied","Data":"c92792f6d2e7d39a1ce700c1e84f032ce8f7da2035aeca94f08b7a6de4a897ca"} Dec 03 00:51:55 crc kubenswrapper[4912]: I1203 00:51:55.962899 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerStarted","Data":"3bdba24c3b1f97ecd218b9544b51816afbac188563cd580280c0bda1c02495b9"} Dec 03 00:51:55 crc kubenswrapper[4912]: I1203 00:51:55.963161 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:51:56 crc kubenswrapper[4912]: I1203 00:51:56.050213 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.946031396 podStartE2EDuration="5.050181083s" podCreationTimestamp="2025-12-03 00:51:51 +0000 UTC" firstStartedPulling="2025-12-03 00:51:52.206094429 +0000 UTC m=+1697.848114989" lastFinishedPulling="2025-12-03 00:51:55.310244116 +0000 UTC m=+1700.952264676" observedRunningTime="2025-12-03 00:51:56.02971587 +0000 UTC m=+1701.671736430" watchObservedRunningTime="2025-12-03 00:51:56.050181083 +0000 UTC m=+1701.692201643" Dec 03 00:51:56 crc kubenswrapper[4912]: I1203 00:51:56.269465 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 00:51:56 crc kubenswrapper[4912]: I1203 00:51:56.272987 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 00:51:56 crc kubenswrapper[4912]: I1203 00:51:56.277076 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 00:51:56 crc kubenswrapper[4912]: I1203 00:51:56.991716 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.215515 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.215591 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.714732 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.810400 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-config-data\") pod \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.810560 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmk9j\" (UniqueName: \"kubernetes.io/projected/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-kube-api-access-zmk9j\") pod \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.810667 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-scripts\") pod \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.810760 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-combined-ca-bundle\") pod \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\" (UID: \"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9\") " Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.820260 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-scripts" (OuterVolumeSpecName: "scripts") pod "a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9" (UID: "a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.845393 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-kube-api-access-zmk9j" (OuterVolumeSpecName: "kube-api-access-zmk9j") pod "a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9" (UID: "a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9"). InnerVolumeSpecName "kube-api-access-zmk9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.857166 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-config-data" (OuterVolumeSpecName: "config-data") pod "a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9" (UID: "a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.860019 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9" (UID: "a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.914371 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.914772 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.914862 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmk9j\" (UniqueName: \"kubernetes.io/projected/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-kube-api-access-zmk9j\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.914937 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.988698 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-sgfjt" event={"ID":"a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9","Type":"ContainerDied","Data":"f817f9b3435a580a8b72e26f0ad68f863e9df7031f88520a40b1000c40137d54"} Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.989105 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f817f9b3435a580a8b72e26f0ad68f863e9df7031f88520a40b1000c40137d54" Dec 03 00:51:57 crc kubenswrapper[4912]: I1203 00:51:57.988751 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-sgfjt" Dec 03 00:51:58 crc kubenswrapper[4912]: I1203 00:51:58.153559 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:51:58 crc kubenswrapper[4912]: I1203 00:51:58.154166 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-api" containerID="cri-o://dd03310ff4d91964a6172396d8d73845fdb0fad9cced0ecfd0f286800339f81e" gracePeriod=30 Dec 03 00:51:58 crc kubenswrapper[4912]: I1203 00:51:58.154583 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-log" containerID="cri-o://3eab385cfdcd041284bf2e2e409dcaff795e0cd10d1dd1e84413a7b58443bef5" gracePeriod=30 Dec 03 00:51:58 crc kubenswrapper[4912]: I1203 00:51:58.168196 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:51:58 crc kubenswrapper[4912]: I1203 00:51:58.168475 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8105f90c-0397-4120-b23d-a0d3d72f6b01" containerName="nova-scheduler-scheduler" containerID="cri-o://2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424" gracePeriod=30 Dec 03 00:51:58 crc kubenswrapper[4912]: I1203 00:51:58.170477 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.235:8774/\": EOF" Dec 03 00:51:58 crc kubenswrapper[4912]: I1203 00:51:58.179378 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.235:8774/\": EOF" Dec 03 00:51:58 crc kubenswrapper[4912]: I1203 00:51:58.186250 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:51:59 crc kubenswrapper[4912]: I1203 00:51:59.002238 4912 generic.go:334] "Generic (PLEG): container finished" podID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerID="3eab385cfdcd041284bf2e2e409dcaff795e0cd10d1dd1e84413a7b58443bef5" exitCode=143 Dec 03 00:51:59 crc kubenswrapper[4912]: I1203 00:51:59.002359 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de4b39f9-20a8-4205-8131-50877d60d1f5","Type":"ContainerDied","Data":"3eab385cfdcd041284bf2e2e409dcaff795e0cd10d1dd1e84413a7b58443bef5"} Dec 03 00:52:00 crc kubenswrapper[4912]: I1203 00:52:00.013186 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-log" containerID="cri-o://5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d" gracePeriod=30 Dec 03 00:52:00 crc kubenswrapper[4912]: I1203 00:52:00.013220 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-metadata" containerID="cri-o://b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3" gracePeriod=30 Dec 03 00:52:01 crc kubenswrapper[4912]: I1203 00:52:01.026167 4912 generic.go:334] "Generic (PLEG): container finished" podID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerID="5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d" exitCode=143 Dec 03 00:52:01 crc kubenswrapper[4912]: I1203 00:52:01.026225 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ccac98c6-e328-46ea-b277-acf80bcc7dec","Type":"ContainerDied","Data":"5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d"} Dec 03 00:52:01 crc kubenswrapper[4912]: I1203 00:52:01.571898 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:52:01 crc kubenswrapper[4912]: E1203 00:52:01.572376 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.033035 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.044449 4912 generic.go:334] "Generic (PLEG): container finished" podID="8105f90c-0397-4120-b23d-a0d3d72f6b01" containerID="2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424" exitCode=0 Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.044514 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8105f90c-0397-4120-b23d-a0d3d72f6b01","Type":"ContainerDied","Data":"2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424"} Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.044533 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.044558 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8105f90c-0397-4120-b23d-a0d3d72f6b01","Type":"ContainerDied","Data":"a5dd41fead2fd7e1aac2078aaf3e87e8fb9f28be2c5e9ef43b7e9e92389142e8"} Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.044579 4912 scope.go:117] "RemoveContainer" containerID="2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.079450 4912 scope.go:117] "RemoveContainer" containerID="2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424" Dec 03 00:52:02 crc kubenswrapper[4912]: E1203 00:52:02.080167 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424\": container with ID starting with 2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424 not found: ID does not exist" containerID="2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.080238 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424"} err="failed to get container status \"2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424\": rpc error: code = NotFound desc = could not find container \"2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424\": container with ID starting with 2dfe926f336a717cbf5679382102d485091ecfd606170084d6f1d45ddc8be424 not found: ID does not exist" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.144124 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-combined-ca-bundle\") pod \"8105f90c-0397-4120-b23d-a0d3d72f6b01\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.144209 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-config-data\") pod \"8105f90c-0397-4120-b23d-a0d3d72f6b01\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.144471 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7twk\" (UniqueName: \"kubernetes.io/projected/8105f90c-0397-4120-b23d-a0d3d72f6b01-kube-api-access-n7twk\") pod \"8105f90c-0397-4120-b23d-a0d3d72f6b01\" (UID: \"8105f90c-0397-4120-b23d-a0d3d72f6b01\") " Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.151704 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8105f90c-0397-4120-b23d-a0d3d72f6b01-kube-api-access-n7twk" (OuterVolumeSpecName: "kube-api-access-n7twk") pod "8105f90c-0397-4120-b23d-a0d3d72f6b01" (UID: "8105f90c-0397-4120-b23d-a0d3d72f6b01"). InnerVolumeSpecName "kube-api-access-n7twk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.177792 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-config-data" (OuterVolumeSpecName: "config-data") pod "8105f90c-0397-4120-b23d-a0d3d72f6b01" (UID: "8105f90c-0397-4120-b23d-a0d3d72f6b01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.183046 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8105f90c-0397-4120-b23d-a0d3d72f6b01" (UID: "8105f90c-0397-4120-b23d-a0d3d72f6b01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.247524 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.247570 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8105f90c-0397-4120-b23d-a0d3d72f6b01-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.247581 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7twk\" (UniqueName: \"kubernetes.io/projected/8105f90c-0397-4120-b23d-a0d3d72f6b01-kube-api-access-n7twk\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.382987 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.398833 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.420233 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:52:02 crc kubenswrapper[4912]: E1203 00:52:02.420978 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9" containerName="nova-manage" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.421002 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9" containerName="nova-manage" Dec 03 00:52:02 crc kubenswrapper[4912]: E1203 00:52:02.421028 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8105f90c-0397-4120-b23d-a0d3d72f6b01" containerName="nova-scheduler-scheduler" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.421036 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8105f90c-0397-4120-b23d-a0d3d72f6b01" containerName="nova-scheduler-scheduler" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.421283 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9" containerName="nova-manage" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.421305 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="8105f90c-0397-4120-b23d-a0d3d72f6b01" containerName="nova-scheduler-scheduler" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.422281 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.436309 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.443374 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.553965 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh2ml\" (UniqueName: \"kubernetes.io/projected/e6cbac08-d327-4238-8335-a69cfb3f71b4-kube-api-access-nh2ml\") pod \"nova-scheduler-0\" (UID: \"e6cbac08-d327-4238-8335-a69cfb3f71b4\") " pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.554120 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6cbac08-d327-4238-8335-a69cfb3f71b4-config-data\") pod \"nova-scheduler-0\" (UID: \"e6cbac08-d327-4238-8335-a69cfb3f71b4\") " pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.554159 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6cbac08-d327-4238-8335-a69cfb3f71b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e6cbac08-d327-4238-8335-a69cfb3f71b4\") " pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.595250 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8105f90c-0397-4120-b23d-a0d3d72f6b01" path="/var/lib/kubelet/pods/8105f90c-0397-4120-b23d-a0d3d72f6b01/volumes" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.657285 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6cbac08-d327-4238-8335-a69cfb3f71b4-config-data\") pod \"nova-scheduler-0\" (UID: \"e6cbac08-d327-4238-8335-a69cfb3f71b4\") " pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.657393 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6cbac08-d327-4238-8335-a69cfb3f71b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e6cbac08-d327-4238-8335-a69cfb3f71b4\") " pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.657586 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh2ml\" (UniqueName: \"kubernetes.io/projected/e6cbac08-d327-4238-8335-a69cfb3f71b4-kube-api-access-nh2ml\") pod \"nova-scheduler-0\" (UID: \"e6cbac08-d327-4238-8335-a69cfb3f71b4\") " pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.661895 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6cbac08-d327-4238-8335-a69cfb3f71b4-config-data\") pod \"nova-scheduler-0\" (UID: \"e6cbac08-d327-4238-8335-a69cfb3f71b4\") " pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.662910 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6cbac08-d327-4238-8335-a69cfb3f71b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e6cbac08-d327-4238-8335-a69cfb3f71b4\") " pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.676114 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh2ml\" (UniqueName: \"kubernetes.io/projected/e6cbac08-d327-4238-8335-a69cfb3f71b4-kube-api-access-nh2ml\") pod \"nova-scheduler-0\" (UID: \"e6cbac08-d327-4238-8335-a69cfb3f71b4\") " pod="openstack/nova-scheduler-0" Dec 03 00:52:02 crc kubenswrapper[4912]: I1203 00:52:02.759208 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.164322 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.231:8775/\": read tcp 10.217.0.2:54408->10.217.0.231:8775: read: connection reset by peer" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.164462 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.231:8775/\": read tcp 10.217.0.2:54418->10.217.0.231:8775: read: connection reset by peer" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.260226 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.644878 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.800369 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vvkn\" (UniqueName: \"kubernetes.io/projected/ccac98c6-e328-46ea-b277-acf80bcc7dec-kube-api-access-7vvkn\") pod \"ccac98c6-e328-46ea-b277-acf80bcc7dec\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.801045 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-config-data\") pod \"ccac98c6-e328-46ea-b277-acf80bcc7dec\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.801105 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-nova-metadata-tls-certs\") pod \"ccac98c6-e328-46ea-b277-acf80bcc7dec\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.801183 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccac98c6-e328-46ea-b277-acf80bcc7dec-logs\") pod \"ccac98c6-e328-46ea-b277-acf80bcc7dec\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.801262 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-combined-ca-bundle\") pod \"ccac98c6-e328-46ea-b277-acf80bcc7dec\" (UID: \"ccac98c6-e328-46ea-b277-acf80bcc7dec\") " Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.806160 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccac98c6-e328-46ea-b277-acf80bcc7dec-logs" (OuterVolumeSpecName: "logs") pod "ccac98c6-e328-46ea-b277-acf80bcc7dec" (UID: "ccac98c6-e328-46ea-b277-acf80bcc7dec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.813133 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccac98c6-e328-46ea-b277-acf80bcc7dec-kube-api-access-7vvkn" (OuterVolumeSpecName: "kube-api-access-7vvkn") pod "ccac98c6-e328-46ea-b277-acf80bcc7dec" (UID: "ccac98c6-e328-46ea-b277-acf80bcc7dec"). InnerVolumeSpecName "kube-api-access-7vvkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.836548 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccac98c6-e328-46ea-b277-acf80bcc7dec" (UID: "ccac98c6-e328-46ea-b277-acf80bcc7dec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.850600 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-config-data" (OuterVolumeSpecName: "config-data") pod "ccac98c6-e328-46ea-b277-acf80bcc7dec" (UID: "ccac98c6-e328-46ea-b277-acf80bcc7dec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.899561 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ccac98c6-e328-46ea-b277-acf80bcc7dec" (UID: "ccac98c6-e328-46ea-b277-acf80bcc7dec"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.904119 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vvkn\" (UniqueName: \"kubernetes.io/projected/ccac98c6-e328-46ea-b277-acf80bcc7dec-kube-api-access-7vvkn\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.904159 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.904170 4912 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.904179 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccac98c6-e328-46ea-b277-acf80bcc7dec-logs\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.904188 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccac98c6-e328-46ea-b277-acf80bcc7dec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:03 crc kubenswrapper[4912]: I1203 00:52:03.984477 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.072868 4912 generic.go:334] "Generic (PLEG): container finished" podID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerID="54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e" exitCode=137 Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.072949 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.072950 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerDied","Data":"54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e"} Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.073027 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"920fd13c-46b2-4fad-a270-c9c900b913f1","Type":"ContainerDied","Data":"c2d3453912a8b7ca1609b55e6d50c64633c05409c8fe7ccaa0a1a9201d984784"} Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.073111 4912 scope.go:117] "RemoveContainer" containerID="54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.076014 4912 generic.go:334] "Generic (PLEG): container finished" podID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerID="b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3" exitCode=0 Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.076085 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ccac98c6-e328-46ea-b277-acf80bcc7dec","Type":"ContainerDied","Data":"b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3"} Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.076116 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ccac98c6-e328-46ea-b277-acf80bcc7dec","Type":"ContainerDied","Data":"2479014a749ed133f576bbe65359cc63c3e06e51dfff406da091079d3e2da61a"} Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.076185 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.085233 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e6cbac08-d327-4238-8335-a69cfb3f71b4","Type":"ContainerStarted","Data":"b2691085fddcf7662a7c595fdec5b3f63fd85350a6cb6176ce74529f87ca36ba"} Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.085346 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e6cbac08-d327-4238-8335-a69cfb3f71b4","Type":"ContainerStarted","Data":"a164a56289f11b787ec11e3be14123c348215cc713512e3883c32bae4fc313c9"} Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.108317 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq692\" (UniqueName: \"kubernetes.io/projected/920fd13c-46b2-4fad-a270-c9c900b913f1-kube-api-access-cq692\") pod \"920fd13c-46b2-4fad-a270-c9c900b913f1\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.108827 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-config-data\") pod \"920fd13c-46b2-4fad-a270-c9c900b913f1\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.108941 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-scripts\") pod \"920fd13c-46b2-4fad-a270-c9c900b913f1\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.109173 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-combined-ca-bundle\") pod \"920fd13c-46b2-4fad-a270-c9c900b913f1\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.122932 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/920fd13c-46b2-4fad-a270-c9c900b913f1-kube-api-access-cq692" (OuterVolumeSpecName: "kube-api-access-cq692") pod "920fd13c-46b2-4fad-a270-c9c900b913f1" (UID: "920fd13c-46b2-4fad-a270-c9c900b913f1"). InnerVolumeSpecName "kube-api-access-cq692". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.127724 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-scripts" (OuterVolumeSpecName: "scripts") pod "920fd13c-46b2-4fad-a270-c9c900b913f1" (UID: "920fd13c-46b2-4fad-a270-c9c900b913f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.128585 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.128556734 podStartE2EDuration="2.128556734s" podCreationTimestamp="2025-12-03 00:52:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:52:04.120939458 +0000 UTC m=+1709.762960038" watchObservedRunningTime="2025-12-03 00:52:04.128556734 +0000 UTC m=+1709.770577294" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.145377 4912 scope.go:117] "RemoveContainer" containerID="03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.208653 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.211790 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq692\" (UniqueName: \"kubernetes.io/projected/920fd13c-46b2-4fad-a270-c9c900b913f1-kube-api-access-cq692\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.211821 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.230323 4912 scope.go:117] "RemoveContainer" containerID="b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.244553 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.254381 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.255109 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-listener" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255133 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-listener" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.255167 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-log" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255176 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-log" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.255187 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-notifier" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255193 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-notifier" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.255219 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-evaluator" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255227 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-evaluator" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.255245 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-metadata" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255252 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-metadata" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.255281 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-api" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255291 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-api" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255549 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-notifier" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255571 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-evaluator" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255592 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-api" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255608 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" containerName="aodh-listener" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255620 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-log" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.255631 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" containerName="nova-metadata-metadata" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.257067 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.259725 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.260503 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.283653 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.290513 4912 scope.go:117] "RemoveContainer" containerID="fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.298292 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-config-data" (OuterVolumeSpecName: "config-data") pod "920fd13c-46b2-4fad-a270-c9c900b913f1" (UID: "920fd13c-46b2-4fad-a270-c9c900b913f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.313735 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "920fd13c-46b2-4fad-a270-c9c900b913f1" (UID: "920fd13c-46b2-4fad-a270-c9c900b913f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.314650 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-combined-ca-bundle\") pod \"920fd13c-46b2-4fad-a270-c9c900b913f1\" (UID: \"920fd13c-46b2-4fad-a270-c9c900b913f1\") " Dec 03 00:52:04 crc kubenswrapper[4912]: W1203 00:52:04.314832 4912 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/920fd13c-46b2-4fad-a270-c9c900b913f1/volumes/kubernetes.io~secret/combined-ca-bundle Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.314876 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "920fd13c-46b2-4fad-a270-c9c900b913f1" (UID: "920fd13c-46b2-4fad-a270-c9c900b913f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.314947 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba93637e-3299-4ead-926f-0e2f84c31df9-config-data\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.314994 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba93637e-3299-4ead-926f-0e2f84c31df9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.315173 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb9rd\" (UniqueName: \"kubernetes.io/projected/ba93637e-3299-4ead-926f-0e2f84c31df9-kube-api-access-nb9rd\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.315235 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba93637e-3299-4ead-926f-0e2f84c31df9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.315724 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba93637e-3299-4ead-926f-0e2f84c31df9-logs\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.315980 4912 scope.go:117] "RemoveContainer" containerID="54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.316057 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.316094 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/920fd13c-46b2-4fad-a270-c9c900b913f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.316622 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e\": container with ID starting with 54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e not found: ID does not exist" containerID="54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.316686 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e"} err="failed to get container status \"54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e\": rpc error: code = NotFound desc = could not find container \"54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e\": container with ID starting with 54b8b479c05c5782b634de84cbd7d124c81e463de968bcb03d538c2759884d1e not found: ID does not exist" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.316720 4912 scope.go:117] "RemoveContainer" containerID="03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.320094 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127\": container with ID starting with 03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127 not found: ID does not exist" containerID="03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.320552 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127"} err="failed to get container status \"03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127\": rpc error: code = NotFound desc = could not find container \"03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127\": container with ID starting with 03449efc74557fd60f166e608f7c5a2dbeb6bfc7f508e2afa6867912af2d0127 not found: ID does not exist" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.320754 4912 scope.go:117] "RemoveContainer" containerID="b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.321167 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1\": container with ID starting with b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1 not found: ID does not exist" containerID="b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.321203 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1"} err="failed to get container status \"b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1\": rpc error: code = NotFound desc = could not find container \"b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1\": container with ID starting with b4fe9b138004e9fd0db52115dce2f4da09897f24e667ff8dbe903e7700cc70e1 not found: ID does not exist" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.322339 4912 scope.go:117] "RemoveContainer" containerID="fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.322860 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f\": container with ID starting with fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f not found: ID does not exist" containerID="fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.322913 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f"} err="failed to get container status \"fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f\": rpc error: code = NotFound desc = could not find container \"fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f\": container with ID starting with fe19b34d910dd301658282e9eee7ce91d2db1fad0b7279434fdca7120d11695f not found: ID does not exist" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.322942 4912 scope.go:117] "RemoveContainer" containerID="b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.354576 4912 scope.go:117] "RemoveContainer" containerID="5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.377138 4912 scope.go:117] "RemoveContainer" containerID="b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.378028 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3\": container with ID starting with b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3 not found: ID does not exist" containerID="b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.378373 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3"} err="failed to get container status \"b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3\": rpc error: code = NotFound desc = could not find container \"b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3\": container with ID starting with b986cc6cf38229d1731bff1a04a4a84ea67cb115b9724ebad92882104735fcc3 not found: ID does not exist" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.378638 4912 scope.go:117] "RemoveContainer" containerID="5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d" Dec 03 00:52:04 crc kubenswrapper[4912]: E1203 00:52:04.379321 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d\": container with ID starting with 5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d not found: ID does not exist" containerID="5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.379390 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d"} err="failed to get container status \"5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d\": rpc error: code = NotFound desc = could not find container \"5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d\": container with ID starting with 5d605f24289ba7ca7c310f5e051ece505a36e5631987a62eba8c2f5cf0d1b91d not found: ID does not exist" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.419407 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba93637e-3299-4ead-926f-0e2f84c31df9-logs\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.419544 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba93637e-3299-4ead-926f-0e2f84c31df9-config-data\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.419614 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba93637e-3299-4ead-926f-0e2f84c31df9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.419690 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb9rd\" (UniqueName: \"kubernetes.io/projected/ba93637e-3299-4ead-926f-0e2f84c31df9-kube-api-access-nb9rd\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.419720 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba93637e-3299-4ead-926f-0e2f84c31df9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.420754 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba93637e-3299-4ead-926f-0e2f84c31df9-logs\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.424160 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ba93637e-3299-4ead-926f-0e2f84c31df9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.424282 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.426527 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba93637e-3299-4ead-926f-0e2f84c31df9-config-data\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.427336 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba93637e-3299-4ead-926f-0e2f84c31df9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.438772 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.441037 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb9rd\" (UniqueName: \"kubernetes.io/projected/ba93637e-3299-4ead-926f-0e2f84c31df9-kube-api-access-nb9rd\") pod \"nova-metadata-0\" (UID: \"ba93637e-3299-4ead-926f-0e2f84c31df9\") " pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.488359 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.491885 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.498134 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.498495 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.499604 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.509757 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tnsxp" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.516640 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.529969 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.586964 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.602809 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="920fd13c-46b2-4fad-a270-c9c900b913f1" path="/var/lib/kubelet/pods/920fd13c-46b2-4fad-a270-c9c900b913f1/volumes" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.604329 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccac98c6-e328-46ea-b277-acf80bcc7dec" path="/var/lib/kubelet/pods/ccac98c6-e328-46ea-b277-acf80bcc7dec/volumes" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.624632 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-config-data\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.625056 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-internal-tls-certs\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.625224 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl7mq\" (UniqueName: \"kubernetes.io/projected/ddca249a-5151-469a-b9c0-34c5ad0ecc86-kube-api-access-dl7mq\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.625344 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-public-tls-certs\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.625527 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-scripts\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.625643 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.751484 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-config-data\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.751625 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-internal-tls-certs\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.751717 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl7mq\" (UniqueName: \"kubernetes.io/projected/ddca249a-5151-469a-b9c0-34c5ad0ecc86-kube-api-access-dl7mq\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.751765 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-public-tls-certs\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.751802 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-scripts\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.751818 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.783478 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-internal-tls-certs\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.786667 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-public-tls-certs\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.793537 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-config-data\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.796165 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.796477 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-scripts\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.822086 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl7mq\" (UniqueName: \"kubernetes.io/projected/ddca249a-5151-469a-b9c0-34c5ad0ecc86-kube-api-access-dl7mq\") pod \"aodh-0\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " pod="openstack/aodh-0" Dec 03 00:52:04 crc kubenswrapper[4912]: I1203 00:52:04.824026 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.109682 4912 generic.go:334] "Generic (PLEG): container finished" podID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerID="dd03310ff4d91964a6172396d8d73845fdb0fad9cced0ecfd0f286800339f81e" exitCode=0 Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.110089 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de4b39f9-20a8-4205-8131-50877d60d1f5","Type":"ContainerDied","Data":"dd03310ff4d91964a6172396d8d73845fdb0fad9cced0ecfd0f286800339f81e"} Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.385910 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.551987 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.556237 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.598967 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.687054 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de4b39f9-20a8-4205-8131-50877d60d1f5-logs\") pod \"de4b39f9-20a8-4205-8131-50877d60d1f5\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.687697 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b74mv\" (UniqueName: \"kubernetes.io/projected/de4b39f9-20a8-4205-8131-50877d60d1f5-kube-api-access-b74mv\") pod \"de4b39f9-20a8-4205-8131-50877d60d1f5\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.687747 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-combined-ca-bundle\") pod \"de4b39f9-20a8-4205-8131-50877d60d1f5\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.687864 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-internal-tls-certs\") pod \"de4b39f9-20a8-4205-8131-50877d60d1f5\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.687996 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-public-tls-certs\") pod \"de4b39f9-20a8-4205-8131-50877d60d1f5\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.687995 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de4b39f9-20a8-4205-8131-50877d60d1f5-logs" (OuterVolumeSpecName: "logs") pod "de4b39f9-20a8-4205-8131-50877d60d1f5" (UID: "de4b39f9-20a8-4205-8131-50877d60d1f5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.688028 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-config-data\") pod \"de4b39f9-20a8-4205-8131-50877d60d1f5\" (UID: \"de4b39f9-20a8-4205-8131-50877d60d1f5\") " Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.689604 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de4b39f9-20a8-4205-8131-50877d60d1f5-logs\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.700111 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de4b39f9-20a8-4205-8131-50877d60d1f5-kube-api-access-b74mv" (OuterVolumeSpecName: "kube-api-access-b74mv") pod "de4b39f9-20a8-4205-8131-50877d60d1f5" (UID: "de4b39f9-20a8-4205-8131-50877d60d1f5"). InnerVolumeSpecName "kube-api-access-b74mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.738632 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-config-data" (OuterVolumeSpecName: "config-data") pod "de4b39f9-20a8-4205-8131-50877d60d1f5" (UID: "de4b39f9-20a8-4205-8131-50877d60d1f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.762703 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de4b39f9-20a8-4205-8131-50877d60d1f5" (UID: "de4b39f9-20a8-4205-8131-50877d60d1f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.779103 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "de4b39f9-20a8-4205-8131-50877d60d1f5" (UID: "de4b39f9-20a8-4205-8131-50877d60d1f5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.793104 4912 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.795702 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.795728 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b74mv\" (UniqueName: \"kubernetes.io/projected/de4b39f9-20a8-4205-8131-50877d60d1f5-kube-api-access-b74mv\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.795744 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.800480 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "de4b39f9-20a8-4205-8131-50877d60d1f5" (UID: "de4b39f9-20a8-4205-8131-50877d60d1f5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:05 crc kubenswrapper[4912]: I1203 00:52:05.899785 4912 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b39f9-20a8-4205-8131-50877d60d1f5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.136966 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ba93637e-3299-4ead-926f-0e2f84c31df9","Type":"ContainerStarted","Data":"5e8d8753ffcf0453b4d42637b8a311105bdd9c293d159967c319acf861360a0d"} Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.138482 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ba93637e-3299-4ead-926f-0e2f84c31df9","Type":"ContainerStarted","Data":"f16eea754de6e1c5aee991c953637b48f8d1e9c9eae62f811d0904ed8fcc552a"} Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.138502 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ba93637e-3299-4ead-926f-0e2f84c31df9","Type":"ContainerStarted","Data":"06f4942720b1040ef96f80a495d37d766bb492256866586d820ee430624463f5"} Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.144097 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerStarted","Data":"95a69893430f83bb8d0ba6688777b1b0e90e8b49ceae110ae446e1cd0f0c58ab"} Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.147556 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de4b39f9-20a8-4205-8131-50877d60d1f5","Type":"ContainerDied","Data":"1d5acafe73aed7932fc82fab964bdc93f99d4f546ee638ab14db2d2fcb52664d"} Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.147613 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.147630 4912 scope.go:117] "RemoveContainer" containerID="dd03310ff4d91964a6172396d8d73845fdb0fad9cced0ecfd0f286800339f81e" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.168071 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.168045684 podStartE2EDuration="2.168045684s" podCreationTimestamp="2025-12-03 00:52:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:52:06.166299547 +0000 UTC m=+1711.808320107" watchObservedRunningTime="2025-12-03 00:52:06.168045684 +0000 UTC m=+1711.810066234" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.189735 4912 scope.go:117] "RemoveContainer" containerID="3eab385cfdcd041284bf2e2e409dcaff795e0cd10d1dd1e84413a7b58443bef5" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.230764 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.250960 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.270073 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 00:52:06 crc kubenswrapper[4912]: E1203 00:52:06.270800 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-api" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.270824 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-api" Dec 03 00:52:06 crc kubenswrapper[4912]: E1203 00:52:06.270876 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-log" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.270884 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-log" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.271109 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-log" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.271125 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" containerName="nova-api-api" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.273052 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.279365 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.281798 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.281724 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.286710 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.314594 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-internal-tls-certs\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.315059 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.315244 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlzmb\" (UniqueName: \"kubernetes.io/projected/23c2e51b-07f1-4b86-bc02-29969db68f38-kube-api-access-hlzmb\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.315391 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23c2e51b-07f1-4b86-bc02-29969db68f38-logs\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.315682 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-config-data\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.315900 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-public-tls-certs\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.418346 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.418532 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlzmb\" (UniqueName: \"kubernetes.io/projected/23c2e51b-07f1-4b86-bc02-29969db68f38-kube-api-access-hlzmb\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.418594 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23c2e51b-07f1-4b86-bc02-29969db68f38-logs\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.418645 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-config-data\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.418751 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-public-tls-certs\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.418775 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-internal-tls-certs\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.419715 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23c2e51b-07f1-4b86-bc02-29969db68f38-logs\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.426130 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-config-data\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.426336 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-internal-tls-certs\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.426580 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.429959 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23c2e51b-07f1-4b86-bc02-29969db68f38-public-tls-certs\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.437807 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlzmb\" (UniqueName: \"kubernetes.io/projected/23c2e51b-07f1-4b86-bc02-29969db68f38-kube-api-access-hlzmb\") pod \"nova-api-0\" (UID: \"23c2e51b-07f1-4b86-bc02-29969db68f38\") " pod="openstack/nova-api-0" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.598367 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de4b39f9-20a8-4205-8131-50877d60d1f5" path="/var/lib/kubelet/pods/de4b39f9-20a8-4205-8131-50877d60d1f5/volumes" Dec 03 00:52:06 crc kubenswrapper[4912]: I1203 00:52:06.636049 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 00:52:07 crc kubenswrapper[4912]: I1203 00:52:07.164481 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerStarted","Data":"391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5"} Dec 03 00:52:07 crc kubenswrapper[4912]: I1203 00:52:07.218705 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 00:52:07 crc kubenswrapper[4912]: W1203 00:52:07.228544 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23c2e51b_07f1_4b86_bc02_29969db68f38.slice/crio-cf2c12effd1b092e834db49bb600a10a0883c182cb5341b2bbed2c80e12f822b WatchSource:0}: Error finding container cf2c12effd1b092e834db49bb600a10a0883c182cb5341b2bbed2c80e12f822b: Status 404 returned error can't find the container with id cf2c12effd1b092e834db49bb600a10a0883c182cb5341b2bbed2c80e12f822b Dec 03 00:52:07 crc kubenswrapper[4912]: I1203 00:52:07.759638 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 00:52:08 crc kubenswrapper[4912]: I1203 00:52:08.184951 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23c2e51b-07f1-4b86-bc02-29969db68f38","Type":"ContainerStarted","Data":"f196bee62d2e9fde48bb2cb2812c2225ef8204f7bcecb731866ae28472461ba5"} Dec 03 00:52:08 crc kubenswrapper[4912]: I1203 00:52:08.185393 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23c2e51b-07f1-4b86-bc02-29969db68f38","Type":"ContainerStarted","Data":"c9aab8963540b6a00816624eab82e8c8abc1dd478861ffbbd2c05ff48bd71d5f"} Dec 03 00:52:08 crc kubenswrapper[4912]: I1203 00:52:08.185409 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23c2e51b-07f1-4b86-bc02-29969db68f38","Type":"ContainerStarted","Data":"cf2c12effd1b092e834db49bb600a10a0883c182cb5341b2bbed2c80e12f822b"} Dec 03 00:52:08 crc kubenswrapper[4912]: I1203 00:52:08.189251 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerStarted","Data":"2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a"} Dec 03 00:52:08 crc kubenswrapper[4912]: I1203 00:52:08.189315 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerStarted","Data":"710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75"} Dec 03 00:52:08 crc kubenswrapper[4912]: I1203 00:52:08.225030 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.225001905 podStartE2EDuration="2.225001905s" podCreationTimestamp="2025-12-03 00:52:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:52:08.215712635 +0000 UTC m=+1713.857733195" watchObservedRunningTime="2025-12-03 00:52:08.225001905 +0000 UTC m=+1713.867022465" Dec 03 00:52:09 crc kubenswrapper[4912]: I1203 00:52:09.204836 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerStarted","Data":"f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd"} Dec 03 00:52:09 crc kubenswrapper[4912]: I1203 00:52:09.255513 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.070562205 podStartE2EDuration="5.255479859s" podCreationTimestamp="2025-12-03 00:52:04 +0000 UTC" firstStartedPulling="2025-12-03 00:52:05.555949715 +0000 UTC m=+1711.197970275" lastFinishedPulling="2025-12-03 00:52:08.740867369 +0000 UTC m=+1714.382887929" observedRunningTime="2025-12-03 00:52:09.233287191 +0000 UTC m=+1714.875307761" watchObservedRunningTime="2025-12-03 00:52:09.255479859 +0000 UTC m=+1714.897500419" Dec 03 00:52:09 crc kubenswrapper[4912]: I1203 00:52:09.588486 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 00:52:09 crc kubenswrapper[4912]: I1203 00:52:09.588545 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 00:52:12 crc kubenswrapper[4912]: I1203 00:52:12.573472 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:52:12 crc kubenswrapper[4912]: E1203 00:52:12.574458 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:52:12 crc kubenswrapper[4912]: I1203 00:52:12.759797 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 00:52:12 crc kubenswrapper[4912]: I1203 00:52:12.809621 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 00:52:13 crc kubenswrapper[4912]: I1203 00:52:13.288246 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 00:52:14 crc kubenswrapper[4912]: I1203 00:52:14.590454 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 00:52:14 crc kubenswrapper[4912]: I1203 00:52:14.590875 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 00:52:15 crc kubenswrapper[4912]: I1203 00:52:15.610678 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ba93637e-3299-4ead-926f-0e2f84c31df9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.239:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 00:52:15 crc kubenswrapper[4912]: I1203 00:52:15.610678 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ba93637e-3299-4ead-926f-0e2f84c31df9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.239:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 00:52:16 crc kubenswrapper[4912]: I1203 00:52:16.636615 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 00:52:16 crc kubenswrapper[4912]: I1203 00:52:16.637165 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 00:52:17 crc kubenswrapper[4912]: I1203 00:52:17.657662 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23c2e51b-07f1-4b86-bc02-29969db68f38" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.241:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 00:52:17 crc kubenswrapper[4912]: I1203 00:52:17.658024 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23c2e51b-07f1-4b86-bc02-29969db68f38" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.241:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 00:52:21 crc kubenswrapper[4912]: I1203 00:52:21.627639 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 00:52:24 crc kubenswrapper[4912]: I1203 00:52:24.582747 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:52:24 crc kubenswrapper[4912]: E1203 00:52:24.584730 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:52:24 crc kubenswrapper[4912]: I1203 00:52:24.605945 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 00:52:24 crc kubenswrapper[4912]: I1203 00:52:24.608261 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 00:52:24 crc kubenswrapper[4912]: I1203 00:52:24.618248 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 00:52:25 crc kubenswrapper[4912]: I1203 00:52:25.463993 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 00:52:26 crc kubenswrapper[4912]: I1203 00:52:26.644395 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 00:52:26 crc kubenswrapper[4912]: I1203 00:52:26.645163 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 00:52:26 crc kubenswrapper[4912]: I1203 00:52:26.647535 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 00:52:26 crc kubenswrapper[4912]: I1203 00:52:26.660201 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 00:52:26 crc kubenswrapper[4912]: I1203 00:52:26.869050 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 00:52:26 crc kubenswrapper[4912]: I1203 00:52:26.869328 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d4611c7f-5acd-4fb6-a984-ef56f9feeb72" containerName="kube-state-metrics" containerID="cri-o://081242bfe133d709203bcb84c3ce63df276a39cd8480362677e97e0354cd4a75" gracePeriod=30 Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.033238 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.033661 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="6af2587c-a78d-4d05-8ee7-137bdfa8c9a1" containerName="mysqld-exporter" containerID="cri-o://740a3af8062d26930e27ad3a6b7b2a65f5d8cffa7149732694da91302c905cf9" gracePeriod=30 Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.476632 4912 generic.go:334] "Generic (PLEG): container finished" podID="6af2587c-a78d-4d05-8ee7-137bdfa8c9a1" containerID="740a3af8062d26930e27ad3a6b7b2a65f5d8cffa7149732694da91302c905cf9" exitCode=2 Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.477163 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1","Type":"ContainerDied","Data":"740a3af8062d26930e27ad3a6b7b2a65f5d8cffa7149732694da91302c905cf9"} Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.479808 4912 generic.go:334] "Generic (PLEG): container finished" podID="d4611c7f-5acd-4fb6-a984-ef56f9feeb72" containerID="081242bfe133d709203bcb84c3ce63df276a39cd8480362677e97e0354cd4a75" exitCode=2 Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.479925 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d4611c7f-5acd-4fb6-a984-ef56f9feeb72","Type":"ContainerDied","Data":"081242bfe133d709203bcb84c3ce63df276a39cd8480362677e97e0354cd4a75"} Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.479998 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d4611c7f-5acd-4fb6-a984-ef56f9feeb72","Type":"ContainerDied","Data":"8cf0489ab34dad7a912e34bf697b4521a5bc5481a8db469df0f5c91ecdee2c89"} Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.480014 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cf0489ab34dad7a912e34bf697b4521a5bc5481a8db469df0f5c91ecdee2c89" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.483451 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.498038 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.530221 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.606224 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcntt\" (UniqueName: \"kubernetes.io/projected/d4611c7f-5acd-4fb6-a984-ef56f9feeb72-kube-api-access-hcntt\") pod \"d4611c7f-5acd-4fb6-a984-ef56f9feeb72\" (UID: \"d4611c7f-5acd-4fb6-a984-ef56f9feeb72\") " Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.628541 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4611c7f-5acd-4fb6-a984-ef56f9feeb72-kube-api-access-hcntt" (OuterVolumeSpecName: "kube-api-access-hcntt") pod "d4611c7f-5acd-4fb6-a984-ef56f9feeb72" (UID: "d4611c7f-5acd-4fb6-a984-ef56f9feeb72"). InnerVolumeSpecName "kube-api-access-hcntt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.698227 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.725754 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcntt\" (UniqueName: \"kubernetes.io/projected/d4611c7f-5acd-4fb6-a984-ef56f9feeb72-kube-api-access-hcntt\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.827683 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-config-data\") pod \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.827781 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-combined-ca-bundle\") pod \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.828192 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cgq9\" (UniqueName: \"kubernetes.io/projected/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-kube-api-access-8cgq9\") pod \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\" (UID: \"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1\") " Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.833191 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-kube-api-access-8cgq9" (OuterVolumeSpecName: "kube-api-access-8cgq9") pod "6af2587c-a78d-4d05-8ee7-137bdfa8c9a1" (UID: "6af2587c-a78d-4d05-8ee7-137bdfa8c9a1"). InnerVolumeSpecName "kube-api-access-8cgq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.862330 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6af2587c-a78d-4d05-8ee7-137bdfa8c9a1" (UID: "6af2587c-a78d-4d05-8ee7-137bdfa8c9a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.890847 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-config-data" (OuterVolumeSpecName: "config-data") pod "6af2587c-a78d-4d05-8ee7-137bdfa8c9a1" (UID: "6af2587c-a78d-4d05-8ee7-137bdfa8c9a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.931976 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.932029 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:27 crc kubenswrapper[4912]: I1203 00:52:27.932052 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cgq9\" (UniqueName: \"kubernetes.io/projected/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1-kube-api-access-8cgq9\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.494782 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.494819 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.494878 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"6af2587c-a78d-4d05-8ee7-137bdfa8c9a1","Type":"ContainerDied","Data":"db2ded104971550f60bc9edcf0fdc1aedb6c4f95bfeca43fe8e5119e4d1d0c12"} Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.496316 4912 scope.go:117] "RemoveContainer" containerID="740a3af8062d26930e27ad3a6b7b2a65f5d8cffa7149732694da91302c905cf9" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.550365 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.596887 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.609934 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.647896 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.670554 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 00:52:28 crc kubenswrapper[4912]: E1203 00:52:28.671408 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6af2587c-a78d-4d05-8ee7-137bdfa8c9a1" containerName="mysqld-exporter" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.671432 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="6af2587c-a78d-4d05-8ee7-137bdfa8c9a1" containerName="mysqld-exporter" Dec 03 00:52:28 crc kubenswrapper[4912]: E1203 00:52:28.671495 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4611c7f-5acd-4fb6-a984-ef56f9feeb72" containerName="kube-state-metrics" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.671503 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4611c7f-5acd-4fb6-a984-ef56f9feeb72" containerName="kube-state-metrics" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.671764 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4611c7f-5acd-4fb6-a984-ef56f9feeb72" containerName="kube-state-metrics" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.671803 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="6af2587c-a78d-4d05-8ee7-137bdfa8c9a1" containerName="mysqld-exporter" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.672945 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.675040 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.676595 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.691227 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.706648 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.710849 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.713422 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.715217 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.730369 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.764379 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610be762-7f1b-4435-8764-eeac647f901c-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.764629 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6rx2\" (UniqueName: \"kubernetes.io/projected/610be762-7f1b-4435-8764-eeac647f901c-kube-api-access-x6rx2\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.764670 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610be762-7f1b-4435-8764-eeac647f901c-config-data\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.764704 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/610be762-7f1b-4435-8764-eeac647f901c-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.867413 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.867534 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.867758 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjzk8\" (UniqueName: \"kubernetes.io/projected/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-kube-api-access-bjzk8\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.867951 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610be762-7f1b-4435-8764-eeac647f901c-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.867995 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.868195 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6rx2\" (UniqueName: \"kubernetes.io/projected/610be762-7f1b-4435-8764-eeac647f901c-kube-api-access-x6rx2\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.868233 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/610be762-7f1b-4435-8764-eeac647f901c-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.868257 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610be762-7f1b-4435-8764-eeac647f901c-config-data\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.873074 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610be762-7f1b-4435-8764-eeac647f901c-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.873368 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610be762-7f1b-4435-8764-eeac647f901c-config-data\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.874499 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/610be762-7f1b-4435-8764-eeac647f901c-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.906892 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6rx2\" (UniqueName: \"kubernetes.io/projected/610be762-7f1b-4435-8764-eeac647f901c-kube-api-access-x6rx2\") pod \"mysqld-exporter-0\" (UID: \"610be762-7f1b-4435-8764-eeac647f901c\") " pod="openstack/mysqld-exporter-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.971007 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.971495 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.971648 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.971767 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjzk8\" (UniqueName: \"kubernetes.io/projected/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-kube-api-access-bjzk8\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.976581 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.977133 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.977140 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:28 crc kubenswrapper[4912]: I1203 00:52:28.994705 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjzk8\" (UniqueName: \"kubernetes.io/projected/4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba-kube-api-access-bjzk8\") pod \"kube-state-metrics-0\" (UID: \"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba\") " pod="openstack/kube-state-metrics-0" Dec 03 00:52:29 crc kubenswrapper[4912]: I1203 00:52:29.000546 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 00:52:29 crc kubenswrapper[4912]: I1203 00:52:29.030199 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 00:52:29 crc kubenswrapper[4912]: I1203 00:52:29.811657 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 00:52:29 crc kubenswrapper[4912]: I1203 00:52:29.825029 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 00:52:29 crc kubenswrapper[4912]: I1203 00:52:29.964299 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:52:29 crc kubenswrapper[4912]: I1203 00:52:29.966074 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="sg-core" containerID="cri-o://5bed58584b46e63c99f836c9f9e2b3506135b293d4f3fa1a58851c90997349ba" gracePeriod=30 Dec 03 00:52:29 crc kubenswrapper[4912]: I1203 00:52:29.966223 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="proxy-httpd" containerID="cri-o://3bdba24c3b1f97ecd218b9544b51816afbac188563cd580280c0bda1c02495b9" gracePeriod=30 Dec 03 00:52:29 crc kubenswrapper[4912]: I1203 00:52:29.966358 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="ceilometer-notification-agent" containerID="cri-o://ae28558032828c1ff946132550939e18bcbf094f28fc719fc67fb5d85150bd99" gracePeriod=30 Dec 03 00:52:29 crc kubenswrapper[4912]: I1203 00:52:29.966526 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="ceilometer-central-agent" containerID="cri-o://f37336a7f9a33dd358b2b1672a23e94279c67b8437827f79c8726ac95070ad3e" gracePeriod=30 Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.569811 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"610be762-7f1b-4435-8764-eeac647f901c","Type":"ContainerStarted","Data":"012ef7d48cb3aaeab806fbc62313fffdfab01eca27a23d52c10efdf1f97a69fb"} Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.592459 4912 generic.go:334] "Generic (PLEG): container finished" podID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerID="3bdba24c3b1f97ecd218b9544b51816afbac188563cd580280c0bda1c02495b9" exitCode=0 Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.592500 4912 generic.go:334] "Generic (PLEG): container finished" podID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerID="5bed58584b46e63c99f836c9f9e2b3506135b293d4f3fa1a58851c90997349ba" exitCode=2 Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.592524 4912 generic.go:334] "Generic (PLEG): container finished" podID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerID="f37336a7f9a33dd358b2b1672a23e94279c67b8437827f79c8726ac95070ad3e" exitCode=0 Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.611052 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6af2587c-a78d-4d05-8ee7-137bdfa8c9a1" path="/var/lib/kubelet/pods/6af2587c-a78d-4d05-8ee7-137bdfa8c9a1/volumes" Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.613062 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4611c7f-5acd-4fb6-a984-ef56f9feeb72" path="/var/lib/kubelet/pods/d4611c7f-5acd-4fb6-a984-ef56f9feeb72/volumes" Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.613809 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerDied","Data":"3bdba24c3b1f97ecd218b9544b51816afbac188563cd580280c0bda1c02495b9"} Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.614174 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerDied","Data":"5bed58584b46e63c99f836c9f9e2b3506135b293d4f3fa1a58851c90997349ba"} Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.614256 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerDied","Data":"f37336a7f9a33dd358b2b1672a23e94279c67b8437827f79c8726ac95070ad3e"} Dec 03 00:52:30 crc kubenswrapper[4912]: I1203 00:52:30.614505 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba","Type":"ContainerStarted","Data":"9c375bb06d3894edacf249b66e85f1a0f75dc6ea83a4efde4578e0a1beaa1451"} Dec 03 00:52:31 crc kubenswrapper[4912]: I1203 00:52:31.622247 4912 generic.go:334] "Generic (PLEG): container finished" podID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerID="ae28558032828c1ff946132550939e18bcbf094f28fc719fc67fb5d85150bd99" exitCode=0 Dec 03 00:52:31 crc kubenswrapper[4912]: I1203 00:52:31.622317 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerDied","Data":"ae28558032828c1ff946132550939e18bcbf094f28fc719fc67fb5d85150bd99"} Dec 03 00:52:31 crc kubenswrapper[4912]: I1203 00:52:31.624496 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba","Type":"ContainerStarted","Data":"8b70e11d407bb0aa002e797321f6160f56b57f2fdb57d62628edef765d037ab4"} Dec 03 00:52:31 crc kubenswrapper[4912]: I1203 00:52:31.624614 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 00:52:31 crc kubenswrapper[4912]: I1203 00:52:31.626573 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"610be762-7f1b-4435-8764-eeac647f901c","Type":"ContainerStarted","Data":"f501ceecf84822858076a8348d62d43a12cdcc34116d26350adce9181a990f55"} Dec 03 00:52:31 crc kubenswrapper[4912]: I1203 00:52:31.644950 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.192967599 podStartE2EDuration="3.64493526s" podCreationTimestamp="2025-12-03 00:52:28 +0000 UTC" firstStartedPulling="2025-12-03 00:52:29.817241923 +0000 UTC m=+1735.459262493" lastFinishedPulling="2025-12-03 00:52:30.269209594 +0000 UTC m=+1735.911230154" observedRunningTime="2025-12-03 00:52:31.641814336 +0000 UTC m=+1737.283834896" watchObservedRunningTime="2025-12-03 00:52:31.64493526 +0000 UTC m=+1737.286955820" Dec 03 00:52:31 crc kubenswrapper[4912]: I1203 00:52:31.681538 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=3.111559813 podStartE2EDuration="3.681485535s" podCreationTimestamp="2025-12-03 00:52:28 +0000 UTC" firstStartedPulling="2025-12-03 00:52:29.83268276 +0000 UTC m=+1735.474703320" lastFinishedPulling="2025-12-03 00:52:30.402608482 +0000 UTC m=+1736.044629042" observedRunningTime="2025-12-03 00:52:31.663333126 +0000 UTC m=+1737.305353696" watchObservedRunningTime="2025-12-03 00:52:31.681485535 +0000 UTC m=+1737.323506105" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.299258 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.394345 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-run-httpd\") pod \"a7a044e0-a3f5-462d-b348-20f1382d24dd\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.394988 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6skmw\" (UniqueName: \"kubernetes.io/projected/a7a044e0-a3f5-462d-b348-20f1382d24dd-kube-api-access-6skmw\") pod \"a7a044e0-a3f5-462d-b348-20f1382d24dd\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.395148 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-combined-ca-bundle\") pod \"a7a044e0-a3f5-462d-b348-20f1382d24dd\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.395258 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-config-data\") pod \"a7a044e0-a3f5-462d-b348-20f1382d24dd\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.395362 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-sg-core-conf-yaml\") pod \"a7a044e0-a3f5-462d-b348-20f1382d24dd\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.394876 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a7a044e0-a3f5-462d-b348-20f1382d24dd" (UID: "a7a044e0-a3f5-462d-b348-20f1382d24dd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.397279 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-log-httpd\") pod \"a7a044e0-a3f5-462d-b348-20f1382d24dd\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.397340 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-scripts\") pod \"a7a044e0-a3f5-462d-b348-20f1382d24dd\" (UID: \"a7a044e0-a3f5-462d-b348-20f1382d24dd\") " Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.398474 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.398590 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a7a044e0-a3f5-462d-b348-20f1382d24dd" (UID: "a7a044e0-a3f5-462d-b348-20f1382d24dd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.406994 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a044e0-a3f5-462d-b348-20f1382d24dd-kube-api-access-6skmw" (OuterVolumeSpecName: "kube-api-access-6skmw") pod "a7a044e0-a3f5-462d-b348-20f1382d24dd" (UID: "a7a044e0-a3f5-462d-b348-20f1382d24dd"). InnerVolumeSpecName "kube-api-access-6skmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.419597 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-scripts" (OuterVolumeSpecName: "scripts") pod "a7a044e0-a3f5-462d-b348-20f1382d24dd" (UID: "a7a044e0-a3f5-462d-b348-20f1382d24dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.476553 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a7a044e0-a3f5-462d-b348-20f1382d24dd" (UID: "a7a044e0-a3f5-462d-b348-20f1382d24dd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.501296 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.501581 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6skmw\" (UniqueName: \"kubernetes.io/projected/a7a044e0-a3f5-462d-b348-20f1382d24dd-kube-api-access-6skmw\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.501601 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.501615 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a7a044e0-a3f5-462d-b348-20f1382d24dd-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.506917 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7a044e0-a3f5-462d-b348-20f1382d24dd" (UID: "a7a044e0-a3f5-462d-b348-20f1382d24dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.588624 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-config-data" (OuterVolumeSpecName: "config-data") pod "a7a044e0-a3f5-462d-b348-20f1382d24dd" (UID: "a7a044e0-a3f5-462d-b348-20f1382d24dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.604890 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.604929 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7a044e0-a3f5-462d-b348-20f1382d24dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.644576 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a7a044e0-a3f5-462d-b348-20f1382d24dd","Type":"ContainerDied","Data":"b4c4ee09e7f88aa73f8cd72870bda2a5fc46656364fc3086fd45d706beac1f9b"} Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.644715 4912 scope.go:117] "RemoveContainer" containerID="3bdba24c3b1f97ecd218b9544b51816afbac188563cd580280c0bda1c02495b9" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.644964 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.682421 4912 scope.go:117] "RemoveContainer" containerID="5bed58584b46e63c99f836c9f9e2b3506135b293d4f3fa1a58851c90997349ba" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.683681 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.696799 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.717856 4912 scope.go:117] "RemoveContainer" containerID="ae28558032828c1ff946132550939e18bcbf094f28fc719fc67fb5d85150bd99" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.726116 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:52:32 crc kubenswrapper[4912]: E1203 00:52:32.726897 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="ceilometer-central-agent" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.726914 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="ceilometer-central-agent" Dec 03 00:52:32 crc kubenswrapper[4912]: E1203 00:52:32.726947 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="sg-core" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.726955 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="sg-core" Dec 03 00:52:32 crc kubenswrapper[4912]: E1203 00:52:32.726981 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="proxy-httpd" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.726987 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="proxy-httpd" Dec 03 00:52:32 crc kubenswrapper[4912]: E1203 00:52:32.727002 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="ceilometer-notification-agent" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.727009 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="ceilometer-notification-agent" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.727238 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="sg-core" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.727256 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="proxy-httpd" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.727273 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="ceilometer-central-agent" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.727286 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" containerName="ceilometer-notification-agent" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.733917 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.745974 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.746218 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.746440 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.749987 4912 scope.go:117] "RemoveContainer" containerID="f37336a7f9a33dd358b2b1672a23e94279c67b8437827f79c8726ac95070ad3e" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.752937 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.912541 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-config-data\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.912660 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.912914 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.913121 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-scripts\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.913305 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4grfg\" (UniqueName: \"kubernetes.io/projected/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-kube-api-access-4grfg\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.913418 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-log-httpd\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.913716 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:32 crc kubenswrapper[4912]: I1203 00:52:32.913758 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-run-httpd\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.016608 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-config-data\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.016694 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.016753 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.016810 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-scripts\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.016883 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4grfg\" (UniqueName: \"kubernetes.io/projected/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-kube-api-access-4grfg\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.016928 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-log-httpd\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.017033 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.017069 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-run-httpd\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.017549 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-run-httpd\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.017829 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-log-httpd\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.021542 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.022873 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-config-data\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.023039 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.023850 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-scripts\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.026325 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.040647 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4grfg\" (UniqueName: \"kubernetes.io/projected/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-kube-api-access-4grfg\") pod \"ceilometer-0\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.069148 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.632853 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:52:33 crc kubenswrapper[4912]: W1203 00:52:33.643798 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7e4a56a_063e_415f_8a9f_46b9934cdb6a.slice/crio-8eaf3583e1987a82ff3ac84b1180f3248b791754271193189ac4df5d341a003c WatchSource:0}: Error finding container 8eaf3583e1987a82ff3ac84b1180f3248b791754271193189ac4df5d341a003c: Status 404 returned error can't find the container with id 8eaf3583e1987a82ff3ac84b1180f3248b791754271193189ac4df5d341a003c Dec 03 00:52:33 crc kubenswrapper[4912]: I1203 00:52:33.678181 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerStarted","Data":"8eaf3583e1987a82ff3ac84b1180f3248b791754271193189ac4df5d341a003c"} Dec 03 00:52:34 crc kubenswrapper[4912]: I1203 00:52:34.593208 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a044e0-a3f5-462d-b348-20f1382d24dd" path="/var/lib/kubelet/pods/a7a044e0-a3f5-462d-b348-20f1382d24dd/volumes" Dec 03 00:52:34 crc kubenswrapper[4912]: I1203 00:52:34.699974 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerStarted","Data":"fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d"} Dec 03 00:52:35 crc kubenswrapper[4912]: I1203 00:52:35.717756 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerStarted","Data":"cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb"} Dec 03 00:52:36 crc kubenswrapper[4912]: I1203 00:52:36.572175 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:52:36 crc kubenswrapper[4912]: E1203 00:52:36.573018 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:52:36 crc kubenswrapper[4912]: I1203 00:52:36.733130 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerStarted","Data":"8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265"} Dec 03 00:52:38 crc kubenswrapper[4912]: I1203 00:52:38.759929 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerStarted","Data":"a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64"} Dec 03 00:52:38 crc kubenswrapper[4912]: I1203 00:52:38.763752 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:52:38 crc kubenswrapper[4912]: I1203 00:52:38.813000 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5983564 podStartE2EDuration="6.812969067s" podCreationTimestamp="2025-12-03 00:52:32 +0000 UTC" firstStartedPulling="2025-12-03 00:52:33.659600939 +0000 UTC m=+1739.301621489" lastFinishedPulling="2025-12-03 00:52:37.874213596 +0000 UTC m=+1743.516234156" observedRunningTime="2025-12-03 00:52:38.802407442 +0000 UTC m=+1744.444428032" watchObservedRunningTime="2025-12-03 00:52:38.812969067 +0000 UTC m=+1744.454989627" Dec 03 00:52:39 crc kubenswrapper[4912]: I1203 00:52:39.054892 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 00:52:50 crc kubenswrapper[4912]: I1203 00:52:50.582058 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:52:50 crc kubenswrapper[4912]: E1203 00:52:50.583207 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:53:03 crc kubenswrapper[4912]: I1203 00:53:03.082148 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 00:53:03 crc kubenswrapper[4912]: I1203 00:53:03.572263 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:53:03 crc kubenswrapper[4912]: E1203 00:53:03.573518 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.510030 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-kwzvj"] Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.521485 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-kwzvj"] Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.625513 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-69tfm"] Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.627199 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.662700 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-69tfm"] Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.681816 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pllx\" (UniqueName: \"kubernetes.io/projected/11ec043b-b8cb-4971-8fe7-041c85066c43-kube-api-access-5pllx\") pod \"heat-db-sync-69tfm\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.681891 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-combined-ca-bundle\") pod \"heat-db-sync-69tfm\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.682224 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-config-data\") pod \"heat-db-sync-69tfm\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.786703 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pllx\" (UniqueName: \"kubernetes.io/projected/11ec043b-b8cb-4971-8fe7-041c85066c43-kube-api-access-5pllx\") pod \"heat-db-sync-69tfm\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.786827 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-combined-ca-bundle\") pod \"heat-db-sync-69tfm\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.788378 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-config-data\") pod \"heat-db-sync-69tfm\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.795873 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-config-data\") pod \"heat-db-sync-69tfm\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.797153 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-combined-ca-bundle\") pod \"heat-db-sync-69tfm\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.807880 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pllx\" (UniqueName: \"kubernetes.io/projected/11ec043b-b8cb-4971-8fe7-041c85066c43-kube-api-access-5pllx\") pod \"heat-db-sync-69tfm\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:15 crc kubenswrapper[4912]: I1203 00:53:15.976626 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-69tfm" Dec 03 00:53:16 crc kubenswrapper[4912]: I1203 00:53:16.552128 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-69tfm"] Dec 03 00:53:16 crc kubenswrapper[4912]: W1203 00:53:16.564302 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11ec043b_b8cb_4971_8fe7_041c85066c43.slice/crio-809d45581d68227de7b4d93efb580db1b4024525fd993a5ddd1b02d7eca566d3 WatchSource:0}: Error finding container 809d45581d68227de7b4d93efb580db1b4024525fd993a5ddd1b02d7eca566d3: Status 404 returned error can't find the container with id 809d45581d68227de7b4d93efb580db1b4024525fd993a5ddd1b02d7eca566d3 Dec 03 00:53:16 crc kubenswrapper[4912]: I1203 00:53:16.587255 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf24aa53-6e61-4aae-88ed-d5085f10a073" path="/var/lib/kubelet/pods/bf24aa53-6e61-4aae-88ed-d5085f10a073/volumes" Dec 03 00:53:17 crc kubenswrapper[4912]: I1203 00:53:17.275833 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-69tfm" event={"ID":"11ec043b-b8cb-4971-8fe7-041c85066c43","Type":"ContainerStarted","Data":"809d45581d68227de7b4d93efb580db1b4024525fd993a5ddd1b02d7eca566d3"} Dec 03 00:53:17 crc kubenswrapper[4912]: I1203 00:53:17.573188 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:53:17 crc kubenswrapper[4912]: E1203 00:53:17.573545 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:53:17 crc kubenswrapper[4912]: I1203 00:53:17.686944 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 00:53:17 crc kubenswrapper[4912]: I1203 00:53:17.819337 4912 scope.go:117] "RemoveContainer" containerID="081242bfe133d709203bcb84c3ce63df276a39cd8480362677e97e0354cd4a75" Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.026986 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.027355 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="ceilometer-central-agent" containerID="cri-o://fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d" gracePeriod=30 Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.027566 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="sg-core" containerID="cri-o://8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265" gracePeriod=30 Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.027830 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="ceilometer-notification-agent" containerID="cri-o://cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb" gracePeriod=30 Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.027889 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="proxy-httpd" containerID="cri-o://a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64" gracePeriod=30 Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.313972 4912 generic.go:334] "Generic (PLEG): container finished" podID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerID="a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64" exitCode=0 Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.314010 4912 generic.go:334] "Generic (PLEG): container finished" podID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerID="8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265" exitCode=2 Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.314035 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerDied","Data":"a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64"} Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.314068 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerDied","Data":"8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265"} Dec 03 00:53:18 crc kubenswrapper[4912]: I1203 00:53:18.975533 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 00:53:19 crc kubenswrapper[4912]: I1203 00:53:19.337800 4912 generic.go:334] "Generic (PLEG): container finished" podID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerID="fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d" exitCode=0 Dec 03 00:53:19 crc kubenswrapper[4912]: I1203 00:53:19.337859 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerDied","Data":"fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d"} Dec 03 00:53:22 crc kubenswrapper[4912]: I1203 00:53:22.360004 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="38bf4c96-49c7-432f-be55-159b75ed1243" containerName="rabbitmq" containerID="cri-o://c303a0e01c5dd3aed0f794b31d152ce2a5a78a187efb125ebddda24df8be9ae5" gracePeriod=604796 Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.190442 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.336328 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-scripts\") pod \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.336497 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-combined-ca-bundle\") pod \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.336580 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-sg-core-conf-yaml\") pod \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.336663 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-config-data\") pod \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.336685 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-ceilometer-tls-certs\") pod \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.336764 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4grfg\" (UniqueName: \"kubernetes.io/projected/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-kube-api-access-4grfg\") pod \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.336820 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-run-httpd\") pod \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.337004 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-log-httpd\") pod \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\" (UID: \"d7e4a56a-063e-415f-8a9f-46b9934cdb6a\") " Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.338284 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d7e4a56a-063e-415f-8a9f-46b9934cdb6a" (UID: "d7e4a56a-063e-415f-8a9f-46b9934cdb6a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.339297 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d7e4a56a-063e-415f-8a9f-46b9934cdb6a" (UID: "d7e4a56a-063e-415f-8a9f-46b9934cdb6a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.357274 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-scripts" (OuterVolumeSpecName: "scripts") pod "d7e4a56a-063e-415f-8a9f-46b9934cdb6a" (UID: "d7e4a56a-063e-415f-8a9f-46b9934cdb6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.370586 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-kube-api-access-4grfg" (OuterVolumeSpecName: "kube-api-access-4grfg") pod "d7e4a56a-063e-415f-8a9f-46b9934cdb6a" (UID: "d7e4a56a-063e-415f-8a9f-46b9934cdb6a"). InnerVolumeSpecName "kube-api-access-4grfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.387183 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d7e4a56a-063e-415f-8a9f-46b9934cdb6a" (UID: "d7e4a56a-063e-415f-8a9f-46b9934cdb6a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.414851 4912 generic.go:334] "Generic (PLEG): container finished" podID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerID="cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb" exitCode=0 Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.414923 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerDied","Data":"cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb"} Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.414973 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d7e4a56a-063e-415f-8a9f-46b9934cdb6a","Type":"ContainerDied","Data":"8eaf3583e1987a82ff3ac84b1180f3248b791754271193189ac4df5d341a003c"} Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.414994 4912 scope.go:117] "RemoveContainer" containerID="a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.415220 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.440771 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.440817 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.440828 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.440855 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4grfg\" (UniqueName: \"kubernetes.io/projected/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-kube-api-access-4grfg\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.440867 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.460643 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7e4a56a-063e-415f-8a9f-46b9934cdb6a" (UID: "d7e4a56a-063e-415f-8a9f-46b9934cdb6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.473479 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d7e4a56a-063e-415f-8a9f-46b9934cdb6a" (UID: "d7e4a56a-063e-415f-8a9f-46b9934cdb6a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.501498 4912 scope.go:117] "RemoveContainer" containerID="8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.537123 4912 scope.go:117] "RemoveContainer" containerID="cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.540626 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-config-data" (OuterVolumeSpecName: "config-data") pod "d7e4a56a-063e-415f-8a9f-46b9934cdb6a" (UID: "d7e4a56a-063e-415f-8a9f-46b9934cdb6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.546981 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.547040 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.547057 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7e4a56a-063e-415f-8a9f-46b9934cdb6a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.580282 4912 scope.go:117] "RemoveContainer" containerID="fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.620252 4912 scope.go:117] "RemoveContainer" containerID="a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64" Dec 03 00:53:24 crc kubenswrapper[4912]: E1203 00:53:24.621130 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64\": container with ID starting with a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64 not found: ID does not exist" containerID="a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.621175 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64"} err="failed to get container status \"a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64\": rpc error: code = NotFound desc = could not find container \"a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64\": container with ID starting with a3412c675589335821bba98610387dc6efef4fa84d04f8dcbc5d8d0a4036ec64 not found: ID does not exist" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.621200 4912 scope.go:117] "RemoveContainer" containerID="8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265" Dec 03 00:53:24 crc kubenswrapper[4912]: E1203 00:53:24.621620 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265\": container with ID starting with 8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265 not found: ID does not exist" containerID="8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.621637 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265"} err="failed to get container status \"8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265\": rpc error: code = NotFound desc = could not find container \"8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265\": container with ID starting with 8df7023ce6a98010e66ce1aad6e849f900196e180d9a52caaa515afa1e48f265 not found: ID does not exist" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.621654 4912 scope.go:117] "RemoveContainer" containerID="cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb" Dec 03 00:53:24 crc kubenswrapper[4912]: E1203 00:53:24.621890 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb\": container with ID starting with cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb not found: ID does not exist" containerID="cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.621907 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb"} err="failed to get container status \"cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb\": rpc error: code = NotFound desc = could not find container \"cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb\": container with ID starting with cd16af360cf0c6c8115d6c46d3327e4ee16aa15a554660d4d15ae4cc7fdee0bb not found: ID does not exist" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.621919 4912 scope.go:117] "RemoveContainer" containerID="fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d" Dec 03 00:53:24 crc kubenswrapper[4912]: E1203 00:53:24.622112 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d\": container with ID starting with fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d not found: ID does not exist" containerID="fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.622140 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d"} err="failed to get container status \"fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d\": rpc error: code = NotFound desc = could not find container \"fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d\": container with ID starting with fe487b3c8720eb25504d7b5db3032f41bc33124698e210618d9380a056528d2d not found: ID does not exist" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.639040 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" containerName="rabbitmq" containerID="cri-o://f4b259a28c47376f179f856aa118415fa4f7ca51f67b049df6e01b70110ca910" gracePeriod=604795 Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.810954 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.836376 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.866195 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:53:24 crc kubenswrapper[4912]: E1203 00:53:24.866803 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="ceilometer-notification-agent" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.866840 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="ceilometer-notification-agent" Dec 03 00:53:24 crc kubenswrapper[4912]: E1203 00:53:24.866907 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="sg-core" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.866941 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="sg-core" Dec 03 00:53:24 crc kubenswrapper[4912]: E1203 00:53:24.866961 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="ceilometer-central-agent" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.866967 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="ceilometer-central-agent" Dec 03 00:53:24 crc kubenswrapper[4912]: E1203 00:53:24.866997 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="proxy-httpd" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.867005 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="proxy-httpd" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.872930 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="ceilometer-central-agent" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.872993 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="proxy-httpd" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.873028 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="ceilometer-notification-agent" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.873038 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" containerName="sg-core" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.875883 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.879584 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.883610 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.885061 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 00:53:24 crc kubenswrapper[4912]: I1203 00:53:24.904375 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.073992 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.074065 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzsz2\" (UniqueName: \"kubernetes.io/projected/5fd2a312-4b62-4392-baae-306c718f24fe-kube-api-access-nzsz2\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.074110 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-run-httpd\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.074179 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.074204 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-scripts\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.074232 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-log-httpd\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.074247 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.074313 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-config-data\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.176618 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.176668 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-scripts\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.176710 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-log-httpd\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.176733 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.176804 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-config-data\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.176847 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.176877 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzsz2\" (UniqueName: \"kubernetes.io/projected/5fd2a312-4b62-4392-baae-306c718f24fe-kube-api-access-nzsz2\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.176908 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-run-httpd\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.177514 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-run-httpd\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.179971 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-log-httpd\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.183538 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.185159 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-scripts\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.185642 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-config-data\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.199101 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.199641 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.202127 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzsz2\" (UniqueName: \"kubernetes.io/projected/5fd2a312-4b62-4392-baae-306c718f24fe-kube-api-access-nzsz2\") pod \"ceilometer-0\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.222896 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 00:53:25 crc kubenswrapper[4912]: I1203 00:53:25.759038 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 00:53:26 crc kubenswrapper[4912]: I1203 00:53:26.452772 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerStarted","Data":"ccefc08d02a51f3a44a53a1937d41afaae3056296ab10bcea1675d228a920287"} Dec 03 00:53:26 crc kubenswrapper[4912]: I1203 00:53:26.600563 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e4a56a-063e-415f-8a9f-46b9934cdb6a" path="/var/lib/kubelet/pods/d7e4a56a-063e-415f-8a9f-46b9934cdb6a/volumes" Dec 03 00:53:27 crc kubenswrapper[4912]: I1203 00:53:27.842583 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="38bf4c96-49c7-432f-be55-159b75ed1243" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.118:5671: connect: connection refused" Dec 03 00:53:27 crc kubenswrapper[4912]: I1203 00:53:27.995841 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.119:5671: connect: connection refused" Dec 03 00:53:29 crc kubenswrapper[4912]: I1203 00:53:29.512464 4912 generic.go:334] "Generic (PLEG): container finished" podID="38bf4c96-49c7-432f-be55-159b75ed1243" containerID="c303a0e01c5dd3aed0f794b31d152ce2a5a78a187efb125ebddda24df8be9ae5" exitCode=0 Dec 03 00:53:29 crc kubenswrapper[4912]: I1203 00:53:29.512518 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"38bf4c96-49c7-432f-be55-159b75ed1243","Type":"ContainerDied","Data":"c303a0e01c5dd3aed0f794b31d152ce2a5a78a187efb125ebddda24df8be9ae5"} Dec 03 00:53:31 crc kubenswrapper[4912]: I1203 00:53:31.545343 4912 generic.go:334] "Generic (PLEG): container finished" podID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" containerID="f4b259a28c47376f179f856aa118415fa4f7ca51f67b049df6e01b70110ca910" exitCode=0 Dec 03 00:53:31 crc kubenswrapper[4912]: I1203 00:53:31.545484 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2","Type":"ContainerDied","Data":"f4b259a28c47376f179f856aa118415fa4f7ca51f67b049df6e01b70110ca910"} Dec 03 00:53:31 crc kubenswrapper[4912]: I1203 00:53:31.573606 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:53:31 crc kubenswrapper[4912]: E1203 00:53:31.574016 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.196296 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-6br9q"] Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.200193 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.207041 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.223706 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-6br9q"] Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.301401 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.301511 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.301543 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.301828 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.302031 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skh26\" (UniqueName: \"kubernetes.io/projected/c9e8199c-b459-4e55-88d6-f805d2ca430c-kube-api-access-skh26\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.302176 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-config\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.302515 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.405240 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.405314 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.405336 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.405404 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.405494 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skh26\" (UniqueName: \"kubernetes.io/projected/c9e8199c-b459-4e55-88d6-f805d2ca430c-kube-api-access-skh26\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.405546 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-config\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.405640 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.406792 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.407513 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.408166 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.408843 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.413641 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.414864 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-config\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.458659 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skh26\" (UniqueName: \"kubernetes.io/projected/c9e8199c-b459-4e55-88d6-f805d2ca430c-kube-api-access-skh26\") pod \"dnsmasq-dns-5b75489c6f-6br9q\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:36 crc kubenswrapper[4912]: I1203 00:53:36.530629 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.157403 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.174677 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.225723 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-plugins-conf\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.226098 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.226151 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/38bf4c96-49c7-432f-be55-159b75ed1243-pod-info\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.226847 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.226982 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-config-data\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.227026 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-server-conf\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.227084 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-tls\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.227180 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-plugins\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.227790 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-confd\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.227833 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/38bf4c96-49c7-432f-be55-159b75ed1243-erlang-cookie-secret\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.227890 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-erlang-cookie\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.227930 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhsx8\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-kube-api-access-mhsx8\") pod \"38bf4c96-49c7-432f-be55-159b75ed1243\" (UID: \"38bf4c96-49c7-432f-be55-159b75ed1243\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.228935 4912 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.249752 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.257797 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.258110 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.259596 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.283606 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/38bf4c96-49c7-432f-be55-159b75ed1243-pod-info" (OuterVolumeSpecName: "pod-info") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.319537 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38bf4c96-49c7-432f-be55-159b75ed1243-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.326576 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-kube-api-access-mhsx8" (OuterVolumeSpecName: "kube-api-access-mhsx8") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "kube-api-access-mhsx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.329956 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-pod-info\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.329996 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-tls\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.330069 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.330449 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-plugins-conf\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.330555 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-erlang-cookie\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.330628 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-server-conf\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.330670 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-config-data\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.330746 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-erlang-cookie-secret\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.330782 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-confd\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.330859 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j6rf\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-kube-api-access-8j6rf\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.330898 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-plugins\") pod \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\" (UID: \"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2\") " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.331392 4912 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/38bf4c96-49c7-432f-be55-159b75ed1243-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.331404 4912 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.331415 4912 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.331441 4912 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/38bf4c96-49c7-432f-be55-159b75ed1243-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.331450 4912 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.331463 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhsx8\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-kube-api-access-mhsx8\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.331483 4912 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.334533 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-config-data" (OuterVolumeSpecName: "config-data") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.335031 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.345193 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.349102 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.351668 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-pod-info" (OuterVolumeSpecName: "pod-info") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.357171 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.375836 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.392814 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.393054 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-kube-api-access-8j6rf" (OuterVolumeSpecName: "kube-api-access-8j6rf") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "kube-api-access-8j6rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.444240 4912 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.444301 4912 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.444311 4912 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.444367 4912 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.444381 4912 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.444396 4912 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.444965 4912 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.445009 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.445022 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j6rf\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-kube-api-access-8j6rf\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.445056 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-config-data" (OuterVolumeSpecName: "config-data") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.476497 4912 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.477566 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-server-conf" (OuterVolumeSpecName: "server-conf") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.531470 4912 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.556364 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-server-conf" (OuterVolumeSpecName: "server-conf") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.558260 4912 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.558285 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.558314 4912 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.558323 4912 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/38bf4c96-49c7-432f-be55-159b75ed1243-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.558336 4912 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.662867 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "38bf4c96-49c7-432f-be55-159b75ed1243" (UID: "38bf4c96-49c7-432f-be55-159b75ed1243"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.725826 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" (UID: "ef9c2da6-7cb1-48df-b51c-26ced3ec63c2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.743655 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"38bf4c96-49c7-432f-be55-159b75ed1243","Type":"ContainerDied","Data":"369dd507f45a672320ef8d9f416745f870e5836b37b05cb4d774b2cda8fcfa37"} Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.743985 4912 scope.go:117] "RemoveContainer" containerID="c303a0e01c5dd3aed0f794b31d152ce2a5a78a187efb125ebddda24df8be9ae5" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.744052 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.749340 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ef9c2da6-7cb1-48df-b51c-26ced3ec63c2","Type":"ContainerDied","Data":"fbce2ecbf5f2597e41ee011d99013ea45de6413f6420e71f10f2d579b7519515"} Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.749507 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.765767 4912 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.765793 4912 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/38bf4c96-49c7-432f-be55-159b75ed1243-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.800640 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.830009 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.863012 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.887204 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 00:53:37 crc kubenswrapper[4912]: E1203 00:53:37.887864 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" containerName="rabbitmq" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.887884 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" containerName="rabbitmq" Dec 03 00:53:37 crc kubenswrapper[4912]: E1203 00:53:37.887922 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38bf4c96-49c7-432f-be55-159b75ed1243" containerName="rabbitmq" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.887931 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="38bf4c96-49c7-432f-be55-159b75ed1243" containerName="rabbitmq" Dec 03 00:53:37 crc kubenswrapper[4912]: E1203 00:53:37.887952 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" containerName="setup-container" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.887958 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" containerName="setup-container" Dec 03 00:53:37 crc kubenswrapper[4912]: E1203 00:53:37.887971 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38bf4c96-49c7-432f-be55-159b75ed1243" containerName="setup-container" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.887978 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="38bf4c96-49c7-432f-be55-159b75ed1243" containerName="setup-container" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.888212 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" containerName="rabbitmq" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.888235 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="38bf4c96-49c7-432f-be55-159b75ed1243" containerName="rabbitmq" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.889619 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.893359 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.893467 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.893581 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.893685 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-q26vs" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.893756 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.893359 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.893844 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.903362 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.919590 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.933910 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.936559 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.938993 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.939460 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.939594 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.939639 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-klxbd" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.939840 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.948257 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.949742 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.950751 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.970726 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.970907 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.971066 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1435b2ef-93c1-41ac-a2bd-f9af839b7894-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.971146 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.971358 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.971468 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkkz6\" (UniqueName: \"kubernetes.io/projected/1435b2ef-93c1-41ac-a2bd-f9af839b7894-kube-api-access-rkkz6\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.971617 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1435b2ef-93c1-41ac-a2bd-f9af839b7894-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.971722 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1435b2ef-93c1-41ac-a2bd-f9af839b7894-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.971744 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.971831 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1435b2ef-93c1-41ac-a2bd-f9af839b7894-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:37 crc kubenswrapper[4912]: I1203 00:53:37.971939 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1435b2ef-93c1-41ac-a2bd-f9af839b7894-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.074804 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1435b2ef-93c1-41ac-a2bd-f9af839b7894-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.074892 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.074950 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.074976 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb7b3762-41c4-4935-8882-526576167e0a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075087 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075142 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb7b3762-41c4-4935-8882-526576167e0a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075167 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb7b3762-41c4-4935-8882-526576167e0a-config-data\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075216 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb7b3762-41c4-4935-8882-526576167e0a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075251 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1435b2ef-93c1-41ac-a2bd-f9af839b7894-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075331 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075359 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075526 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb7b3762-41c4-4935-8882-526576167e0a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075550 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075639 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075683 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075735 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkkz6\" (UniqueName: \"kubernetes.io/projected/1435b2ef-93c1-41ac-a2bd-f9af839b7894-kube-api-access-rkkz6\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075894 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075911 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075950 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1435b2ef-93c1-41ac-a2bd-f9af839b7894-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.075981 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.076056 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj9rv\" (UniqueName: \"kubernetes.io/projected/eb7b3762-41c4-4935-8882-526576167e0a-kube-api-access-dj9rv\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.076101 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1435b2ef-93c1-41ac-a2bd-f9af839b7894-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.076124 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.076229 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1435b2ef-93c1-41ac-a2bd-f9af839b7894-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.076247 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.076954 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/1435b2ef-93c1-41ac-a2bd-f9af839b7894-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.078944 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/1435b2ef-93c1-41ac-a2bd-f9af839b7894-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.079401 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1435b2ef-93c1-41ac-a2bd-f9af839b7894-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.081318 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.081391 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/1435b2ef-93c1-41ac-a2bd-f9af839b7894-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.083680 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/1435b2ef-93c1-41ac-a2bd-f9af839b7894-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.085832 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/1435b2ef-93c1-41ac-a2bd-f9af839b7894-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.097352 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkkz6\" (UniqueName: \"kubernetes.io/projected/1435b2ef-93c1-41ac-a2bd-f9af839b7894-kube-api-access-rkkz6\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.112136 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"1435b2ef-93c1-41ac-a2bd-f9af839b7894\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179064 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj9rv\" (UniqueName: \"kubernetes.io/projected/eb7b3762-41c4-4935-8882-526576167e0a-kube-api-access-dj9rv\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179198 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179234 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb7b3762-41c4-4935-8882-526576167e0a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179277 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb7b3762-41c4-4935-8882-526576167e0a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179297 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb7b3762-41c4-4935-8882-526576167e0a-config-data\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179327 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb7b3762-41c4-4935-8882-526576167e0a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179355 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb7b3762-41c4-4935-8882-526576167e0a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179376 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179409 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179493 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.179527 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.180487 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/eb7b3762-41c4-4935-8882-526576167e0a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.180938 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/eb7b3762-41c4-4935-8882-526576167e0a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.181120 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/eb7b3762-41c4-4935-8882-526576167e0a-config-data\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.181229 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.181867 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.181967 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.184354 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/eb7b3762-41c4-4935-8882-526576167e0a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.185186 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/eb7b3762-41c4-4935-8882-526576167e0a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.187251 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.189556 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/eb7b3762-41c4-4935-8882-526576167e0a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.203817 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj9rv\" (UniqueName: \"kubernetes.io/projected/eb7b3762-41c4-4935-8882-526576167e0a-kube-api-access-dj9rv\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.217398 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.233533 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"eb7b3762-41c4-4935-8882-526576167e0a\") " pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.281650 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.602654 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38bf4c96-49c7-432f-be55-159b75ed1243" path="/var/lib/kubelet/pods/38bf4c96-49c7-432f-be55-159b75ed1243/volumes" Dec 03 00:53:38 crc kubenswrapper[4912]: I1203 00:53:38.609260 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef9c2da6-7cb1-48df-b51c-26ced3ec63c2" path="/var/lib/kubelet/pods/ef9c2da6-7cb1-48df-b51c-26ced3ec63c2/volumes" Dec 03 00:53:44 crc kubenswrapper[4912]: E1203 00:53:44.414397 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 03 00:53:44 crc kubenswrapper[4912]: E1203 00:53:44.415192 4912 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 03 00:53:44 crc kubenswrapper[4912]: E1203 00:53:44.415356 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5pllx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-69tfm_openstack(11ec043b-b8cb-4971-8fe7-041c85066c43): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:53:44 crc kubenswrapper[4912]: E1203 00:53:44.418008 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-69tfm" podUID="11ec043b-b8cb-4971-8fe7-041c85066c43" Dec 03 00:53:44 crc kubenswrapper[4912]: E1203 00:53:44.854898 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-69tfm" podUID="11ec043b-b8cb-4971-8fe7-041c85066c43" Dec 03 00:53:45 crc kubenswrapper[4912]: I1203 00:53:45.571403 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:53:45 crc kubenswrapper[4912]: E1203 00:53:45.572073 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:53:48 crc kubenswrapper[4912]: E1203 00:53:48.342097 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 03 00:53:48 crc kubenswrapper[4912]: E1203 00:53:48.342514 4912 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 03 00:53:48 crc kubenswrapper[4912]: E1203 00:53:48.342686 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n647h6fhfh5fbh5bh5c7h589h64hfbh5fbh574h679h585hf6h5ch689hf7h67fh675h5d8hd6h6fh57chd9h55dh585h54h59h9ch69hc4h595q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nzsz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(5fd2a312-4b62-4392-baae-306c718f24fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 00:53:48 crc kubenswrapper[4912]: I1203 00:53:48.362943 4912 scope.go:117] "RemoveContainer" containerID="dff02ed2edeeec80c4c16589ce57d80d6f4bff8f20597991b6274a1f56ab981a" Dec 03 00:53:48 crc kubenswrapper[4912]: I1203 00:53:48.553486 4912 scope.go:117] "RemoveContainer" containerID="f4b259a28c47376f179f856aa118415fa4f7ca51f67b049df6e01b70110ca910" Dec 03 00:53:48 crc kubenswrapper[4912]: I1203 00:53:48.600716 4912 scope.go:117] "RemoveContainer" containerID="d601bbaa0a1e7902f9bd53059640462774984a5b945e216fc8163f7c3ae23a3c" Dec 03 00:53:48 crc kubenswrapper[4912]: I1203 00:53:48.854120 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-6br9q"] Dec 03 00:53:48 crc kubenswrapper[4912]: I1203 00:53:48.920828 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" event={"ID":"c9e8199c-b459-4e55-88d6-f805d2ca430c","Type":"ContainerStarted","Data":"7b8a0ace24aea71dbd58eec17da16779fdd76be85e9c589a3d3539a0e23004c2"} Dec 03 00:53:48 crc kubenswrapper[4912]: I1203 00:53:48.992533 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 00:53:49 crc kubenswrapper[4912]: I1203 00:53:49.020981 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 00:53:49 crc kubenswrapper[4912]: I1203 00:53:49.955184 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerStarted","Data":"1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d"} Dec 03 00:53:49 crc kubenswrapper[4912]: I1203 00:53:49.975601 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb7b3762-41c4-4935-8882-526576167e0a","Type":"ContainerStarted","Data":"9611203fb9ae2cbf8342c7e3d71f15966e99d1635fb931c7318dcc881e292fc5"} Dec 03 00:53:49 crc kubenswrapper[4912]: I1203 00:53:49.983752 4912 generic.go:334] "Generic (PLEG): container finished" podID="c9e8199c-b459-4e55-88d6-f805d2ca430c" containerID="0e214b63380071b87afa976450b243427e8eda3f11328ae3f142246e965903dd" exitCode=0 Dec 03 00:53:49 crc kubenswrapper[4912]: I1203 00:53:49.983851 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" event={"ID":"c9e8199c-b459-4e55-88d6-f805d2ca430c","Type":"ContainerDied","Data":"0e214b63380071b87afa976450b243427e8eda3f11328ae3f142246e965903dd"} Dec 03 00:53:49 crc kubenswrapper[4912]: I1203 00:53:49.996939 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1435b2ef-93c1-41ac-a2bd-f9af839b7894","Type":"ContainerStarted","Data":"7d28165c2ad30f3b48b39f6a53cb332401168e368cfff7d7f4477fad620216fb"} Dec 03 00:53:51 crc kubenswrapper[4912]: I1203 00:53:51.019975 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerStarted","Data":"02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4"} Dec 03 00:53:51 crc kubenswrapper[4912]: I1203 00:53:51.027389 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" event={"ID":"c9e8199c-b459-4e55-88d6-f805d2ca430c","Type":"ContainerStarted","Data":"f1dcbc6421bbb80fa82538457d9a82dfb6846533e0830ff975475ccb6c0132c6"} Dec 03 00:53:51 crc kubenswrapper[4912]: I1203 00:53:51.028598 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:51 crc kubenswrapper[4912]: I1203 00:53:51.063331 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" podStartSLOduration=15.06328219 podStartE2EDuration="15.06328219s" podCreationTimestamp="2025-12-03 00:53:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:53:51.05213597 +0000 UTC m=+1816.694156570" watchObservedRunningTime="2025-12-03 00:53:51.06328219 +0000 UTC m=+1816.705302790" Dec 03 00:53:52 crc kubenswrapper[4912]: I1203 00:53:52.044144 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1435b2ef-93c1-41ac-a2bd-f9af839b7894","Type":"ContainerStarted","Data":"6aeda547d26dbbe49ce6f652c700f8f1325e7d5084c650583ca3df31f874f74a"} Dec 03 00:53:52 crc kubenswrapper[4912]: I1203 00:53:52.048487 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb7b3762-41c4-4935-8882-526576167e0a","Type":"ContainerStarted","Data":"5c5da28b70e3c9b4587c70196a56a8efa9e17074769861e73f05fa9fd2b0bfce"} Dec 03 00:53:52 crc kubenswrapper[4912]: E1203 00:53:52.144511 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" Dec 03 00:53:53 crc kubenswrapper[4912]: I1203 00:53:53.064423 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerStarted","Data":"67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74"} Dec 03 00:53:53 crc kubenswrapper[4912]: I1203 00:53:53.066051 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 00:53:53 crc kubenswrapper[4912]: E1203 00:53:53.069850 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" Dec 03 00:53:54 crc kubenswrapper[4912]: E1203 00:53:54.088924 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.533828 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.629895 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-kdbv2"] Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.630222 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" podUID="e700236d-b3e1-40b7-8e95-9e2d799bf0d9" containerName="dnsmasq-dns" containerID="cri-o://3c603903509e566fa9464d11c926c221ae776e6f127fc32229f3c6b5e8e535b6" gracePeriod=10 Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.801966 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf"] Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.808234 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.826209 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf"] Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.898092 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-nb\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.898396 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-openstack-edpm-ipam\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.898565 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-svc\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.898628 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-swift-storage-0\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.898704 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpt78\" (UniqueName: \"kubernetes.io/projected/d819df61-e2e4-4960-8ff8-761a99f21850-kube-api-access-kpt78\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.898773 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-config\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:56 crc kubenswrapper[4912]: I1203 00:53:56.898919 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-sb\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.001318 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-openstack-edpm-ipam\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.001755 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-svc\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.001789 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-swift-storage-0\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.001825 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpt78\" (UniqueName: \"kubernetes.io/projected/d819df61-e2e4-4960-8ff8-761a99f21850-kube-api-access-kpt78\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.001852 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-config\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.001957 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-sb\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.002004 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-nb\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.002994 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-nb\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.003611 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-openstack-edpm-ipam\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.004942 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-svc\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.007476 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-swift-storage-0\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.008051 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-sb\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.013520 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-config\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.043751 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpt78\" (UniqueName: \"kubernetes.io/projected/d819df61-e2e4-4960-8ff8-761a99f21850-kube-api-access-kpt78\") pod \"dnsmasq-dns-5cf7b6cbf7-ctmpf\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.129323 4912 generic.go:334] "Generic (PLEG): container finished" podID="e700236d-b3e1-40b7-8e95-9e2d799bf0d9" containerID="3c603903509e566fa9464d11c926c221ae776e6f127fc32229f3c6b5e8e535b6" exitCode=0 Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.129381 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" event={"ID":"e700236d-b3e1-40b7-8e95-9e2d799bf0d9","Type":"ContainerDied","Data":"3c603903509e566fa9464d11c926c221ae776e6f127fc32229f3c6b5e8e535b6"} Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.129425 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" event={"ID":"e700236d-b3e1-40b7-8e95-9e2d799bf0d9","Type":"ContainerDied","Data":"77689f23753c1140ac6f1fd94eb51cd6ec2c7bf8bbf2f3abe8666be32cc34e1b"} Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.129449 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77689f23753c1140ac6f1fd94eb51cd6ec2c7bf8bbf2f3abe8666be32cc34e1b" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.151644 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.337846 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.412947 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-swift-storage-0\") pod \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.413067 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx9m2\" (UniqueName: \"kubernetes.io/projected/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-kube-api-access-gx9m2\") pod \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.413130 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-svc\") pod \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.413258 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-sb\") pod \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.413299 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-config\") pod \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.413353 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-nb\") pod \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\" (UID: \"e700236d-b3e1-40b7-8e95-9e2d799bf0d9\") " Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.438916 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-kube-api-access-gx9m2" (OuterVolumeSpecName: "kube-api-access-gx9m2") pod "e700236d-b3e1-40b7-8e95-9e2d799bf0d9" (UID: "e700236d-b3e1-40b7-8e95-9e2d799bf0d9"). InnerVolumeSpecName "kube-api-access-gx9m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.482038 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e700236d-b3e1-40b7-8e95-9e2d799bf0d9" (UID: "e700236d-b3e1-40b7-8e95-9e2d799bf0d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.490118 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-config" (OuterVolumeSpecName: "config") pod "e700236d-b3e1-40b7-8e95-9e2d799bf0d9" (UID: "e700236d-b3e1-40b7-8e95-9e2d799bf0d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.495694 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e700236d-b3e1-40b7-8e95-9e2d799bf0d9" (UID: "e700236d-b3e1-40b7-8e95-9e2d799bf0d9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.504172 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e700236d-b3e1-40b7-8e95-9e2d799bf0d9" (UID: "e700236d-b3e1-40b7-8e95-9e2d799bf0d9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.510374 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e700236d-b3e1-40b7-8e95-9e2d799bf0d9" (UID: "e700236d-b3e1-40b7-8e95-9e2d799bf0d9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.519175 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.519231 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.519250 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.519261 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx9m2\" (UniqueName: \"kubernetes.io/projected/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-kube-api-access-gx9m2\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.519270 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.519278 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e700236d-b3e1-40b7-8e95-9e2d799bf0d9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:53:57 crc kubenswrapper[4912]: I1203 00:53:57.730241 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf"] Dec 03 00:53:58 crc kubenswrapper[4912]: I1203 00:53:58.143472 4912 generic.go:334] "Generic (PLEG): container finished" podID="d819df61-e2e4-4960-8ff8-761a99f21850" containerID="300b2fc9e07d213a577e82624659df61c3e7a340b316c0b770cbbad5558a229c" exitCode=0 Dec 03 00:53:58 crc kubenswrapper[4912]: I1203 00:53:58.143584 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" event={"ID":"d819df61-e2e4-4960-8ff8-761a99f21850","Type":"ContainerDied","Data":"300b2fc9e07d213a577e82624659df61c3e7a340b316c0b770cbbad5558a229c"} Dec 03 00:53:58 crc kubenswrapper[4912]: I1203 00:53:58.143979 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-kdbv2" Dec 03 00:53:58 crc kubenswrapper[4912]: I1203 00:53:58.144041 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" event={"ID":"d819df61-e2e4-4960-8ff8-761a99f21850","Type":"ContainerStarted","Data":"c22c68b7d1299076b42d44f10cf2a033f3795bc4b3221d972e171fd4ad45b47a"} Dec 03 00:53:58 crc kubenswrapper[4912]: I1203 00:53:58.363692 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-kdbv2"] Dec 03 00:53:58 crc kubenswrapper[4912]: I1203 00:53:58.374548 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-kdbv2"] Dec 03 00:53:58 crc kubenswrapper[4912]: I1203 00:53:58.586318 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e700236d-b3e1-40b7-8e95-9e2d799bf0d9" path="/var/lib/kubelet/pods/e700236d-b3e1-40b7-8e95-9e2d799bf0d9/volumes" Dec 03 00:53:59 crc kubenswrapper[4912]: I1203 00:53:59.163198 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" event={"ID":"d819df61-e2e4-4960-8ff8-761a99f21850","Type":"ContainerStarted","Data":"3cd1d24a28b305d242e061d5438fee63c1ba6ebc7f164d9027689fc893781107"} Dec 03 00:53:59 crc kubenswrapper[4912]: I1203 00:53:59.163484 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:53:59 crc kubenswrapper[4912]: I1203 00:53:59.220142 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" podStartSLOduration=3.220106027 podStartE2EDuration="3.220106027s" podCreationTimestamp="2025-12-03 00:53:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:53:59.201633909 +0000 UTC m=+1824.843654479" watchObservedRunningTime="2025-12-03 00:53:59.220106027 +0000 UTC m=+1824.862126627" Dec 03 00:54:00 crc kubenswrapper[4912]: I1203 00:54:00.178459 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-69tfm" event={"ID":"11ec043b-b8cb-4971-8fe7-041c85066c43","Type":"ContainerStarted","Data":"03de48ee95541ad02f95324b6119e2e86c6a1ee323173a6bb0d2ae0a3499fc5d"} Dec 03 00:54:00 crc kubenswrapper[4912]: I1203 00:54:00.203639 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-69tfm" podStartSLOduration=2.011577774 podStartE2EDuration="45.203617275s" podCreationTimestamp="2025-12-03 00:53:15 +0000 UTC" firstStartedPulling="2025-12-03 00:53:16.569493669 +0000 UTC m=+1782.211514229" lastFinishedPulling="2025-12-03 00:53:59.76153315 +0000 UTC m=+1825.403553730" observedRunningTime="2025-12-03 00:54:00.201836347 +0000 UTC m=+1825.843856927" watchObservedRunningTime="2025-12-03 00:54:00.203617275 +0000 UTC m=+1825.845637835" Dec 03 00:54:00 crc kubenswrapper[4912]: I1203 00:54:00.572173 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:54:00 crc kubenswrapper[4912]: E1203 00:54:00.572910 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:54:03 crc kubenswrapper[4912]: I1203 00:54:03.276302 4912 generic.go:334] "Generic (PLEG): container finished" podID="11ec043b-b8cb-4971-8fe7-041c85066c43" containerID="03de48ee95541ad02f95324b6119e2e86c6a1ee323173a6bb0d2ae0a3499fc5d" exitCode=0 Dec 03 00:54:03 crc kubenswrapper[4912]: I1203 00:54:03.276579 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-69tfm" event={"ID":"11ec043b-b8cb-4971-8fe7-041c85066c43","Type":"ContainerDied","Data":"03de48ee95541ad02f95324b6119e2e86c6a1ee323173a6bb0d2ae0a3499fc5d"} Dec 03 00:54:04 crc kubenswrapper[4912]: I1203 00:54:04.840660 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-69tfm" Dec 03 00:54:04 crc kubenswrapper[4912]: I1203 00:54:04.956900 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-combined-ca-bundle\") pod \"11ec043b-b8cb-4971-8fe7-041c85066c43\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " Dec 03 00:54:04 crc kubenswrapper[4912]: I1203 00:54:04.957361 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pllx\" (UniqueName: \"kubernetes.io/projected/11ec043b-b8cb-4971-8fe7-041c85066c43-kube-api-access-5pllx\") pod \"11ec043b-b8cb-4971-8fe7-041c85066c43\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " Dec 03 00:54:04 crc kubenswrapper[4912]: I1203 00:54:04.958515 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-config-data\") pod \"11ec043b-b8cb-4971-8fe7-041c85066c43\" (UID: \"11ec043b-b8cb-4971-8fe7-041c85066c43\") " Dec 03 00:54:04 crc kubenswrapper[4912]: I1203 00:54:04.970942 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11ec043b-b8cb-4971-8fe7-041c85066c43-kube-api-access-5pllx" (OuterVolumeSpecName: "kube-api-access-5pllx") pod "11ec043b-b8cb-4971-8fe7-041c85066c43" (UID: "11ec043b-b8cb-4971-8fe7-041c85066c43"). InnerVolumeSpecName "kube-api-access-5pllx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:54:05 crc kubenswrapper[4912]: I1203 00:54:05.008907 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11ec043b-b8cb-4971-8fe7-041c85066c43" (UID: "11ec043b-b8cb-4971-8fe7-041c85066c43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:05 crc kubenswrapper[4912]: I1203 00:54:05.062043 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pllx\" (UniqueName: \"kubernetes.io/projected/11ec043b-b8cb-4971-8fe7-041c85066c43-kube-api-access-5pllx\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:05 crc kubenswrapper[4912]: I1203 00:54:05.062093 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:05 crc kubenswrapper[4912]: I1203 00:54:05.083377 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-config-data" (OuterVolumeSpecName: "config-data") pod "11ec043b-b8cb-4971-8fe7-041c85066c43" (UID: "11ec043b-b8cb-4971-8fe7-041c85066c43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:05 crc kubenswrapper[4912]: I1203 00:54:05.165794 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ec043b-b8cb-4971-8fe7-041c85066c43-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:05 crc kubenswrapper[4912]: I1203 00:54:05.319673 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-69tfm" event={"ID":"11ec043b-b8cb-4971-8fe7-041c85066c43","Type":"ContainerDied","Data":"809d45581d68227de7b4d93efb580db1b4024525fd993a5ddd1b02d7eca566d3"} Dec 03 00:54:05 crc kubenswrapper[4912]: I1203 00:54:05.319765 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="809d45581d68227de7b4d93efb580db1b4024525fd993a5ddd1b02d7eca566d3" Dec 03 00:54:05 crc kubenswrapper[4912]: I1203 00:54:05.319881 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-69tfm" Dec 03 00:54:05 crc kubenswrapper[4912]: I1203 00:54:05.586836 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.331632 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-678985f55c-wftgz"] Dec 03 00:54:06 crc kubenswrapper[4912]: E1203 00:54:06.333977 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ec043b-b8cb-4971-8fe7-041c85066c43" containerName="heat-db-sync" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.334007 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ec043b-b8cb-4971-8fe7-041c85066c43" containerName="heat-db-sync" Dec 03 00:54:06 crc kubenswrapper[4912]: E1203 00:54:06.334055 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e700236d-b3e1-40b7-8e95-9e2d799bf0d9" containerName="init" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.334066 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e700236d-b3e1-40b7-8e95-9e2d799bf0d9" containerName="init" Dec 03 00:54:06 crc kubenswrapper[4912]: E1203 00:54:06.334138 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e700236d-b3e1-40b7-8e95-9e2d799bf0d9" containerName="dnsmasq-dns" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.334151 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e700236d-b3e1-40b7-8e95-9e2d799bf0d9" containerName="dnsmasq-dns" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.334551 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="11ec043b-b8cb-4971-8fe7-041c85066c43" containerName="heat-db-sync" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.334601 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="e700236d-b3e1-40b7-8e95-9e2d799bf0d9" containerName="dnsmasq-dns" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.336159 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.344161 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerStarted","Data":"60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33"} Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.355389 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-678985f55c-wftgz"] Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.406075 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31eb41a-2842-45fe-ba2d-6a84a048723f-combined-ca-bundle\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.406294 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqgmk\" (UniqueName: \"kubernetes.io/projected/e31eb41a-2842-45fe-ba2d-6a84a048723f-kube-api-access-sqgmk\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.409592 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31eb41a-2842-45fe-ba2d-6a84a048723f-config-data\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.409689 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e31eb41a-2842-45fe-ba2d-6a84a048723f-config-data-custom\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.423983 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5cdb6985d5-xgpkb"] Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.426585 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.432857 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.451100138 podStartE2EDuration="42.43283259s" podCreationTimestamp="2025-12-03 00:53:24 +0000 UTC" firstStartedPulling="2025-12-03 00:53:25.766782541 +0000 UTC m=+1791.408803101" lastFinishedPulling="2025-12-03 00:54:05.748514993 +0000 UTC m=+1831.390535553" observedRunningTime="2025-12-03 00:54:06.397830376 +0000 UTC m=+1832.039850956" watchObservedRunningTime="2025-12-03 00:54:06.43283259 +0000 UTC m=+1832.074853150" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.522538 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-combined-ca-bundle\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.522671 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqgmk\" (UniqueName: \"kubernetes.io/projected/e31eb41a-2842-45fe-ba2d-6a84a048723f-kube-api-access-sqgmk\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.522705 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31eb41a-2842-45fe-ba2d-6a84a048723f-config-data\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.522744 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e31eb41a-2842-45fe-ba2d-6a84a048723f-config-data-custom\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.522905 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-config-data\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.523073 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-internal-tls-certs\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.523136 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrhsn\" (UniqueName: \"kubernetes.io/projected/b0e666e2-5db8-409a-9a47-0da8fe895f75-kube-api-access-mrhsn\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.523255 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-public-tls-certs\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.523370 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-config-data-custom\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.523416 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31eb41a-2842-45fe-ba2d-6a84a048723f-combined-ca-bundle\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.534868 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e31eb41a-2842-45fe-ba2d-6a84a048723f-combined-ca-bundle\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.537455 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e31eb41a-2842-45fe-ba2d-6a84a048723f-config-data\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.538275 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e31eb41a-2842-45fe-ba2d-6a84a048723f-config-data-custom\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.569746 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqgmk\" (UniqueName: \"kubernetes.io/projected/e31eb41a-2842-45fe-ba2d-6a84a048723f-kube-api-access-sqgmk\") pod \"heat-engine-678985f55c-wftgz\" (UID: \"e31eb41a-2842-45fe-ba2d-6a84a048723f\") " pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.609804 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5cdb6985d5-xgpkb"] Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.621984 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-56b4585b94-qjwmb"] Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.625108 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.626272 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-config-data\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.627554 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-internal-tls-certs\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.627691 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrhsn\" (UniqueName: \"kubernetes.io/projected/b0e666e2-5db8-409a-9a47-0da8fe895f75-kube-api-access-mrhsn\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.627933 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-public-tls-certs\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.628161 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-config-data-custom\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.628369 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-combined-ca-bundle\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.632208 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-combined-ca-bundle\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.635853 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-public-tls-certs\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.638235 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-config-data\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.638933 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-internal-tls-certs\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.660608 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrhsn\" (UniqueName: \"kubernetes.io/projected/b0e666e2-5db8-409a-9a47-0da8fe895f75-kube-api-access-mrhsn\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.666811 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-56b4585b94-qjwmb"] Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.668081 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.674781 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0e666e2-5db8-409a-9a47-0da8fe895f75-config-data-custom\") pod \"heat-cfnapi-5cdb6985d5-xgpkb\" (UID: \"b0e666e2-5db8-409a-9a47-0da8fe895f75\") " pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.731184 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-internal-tls-certs\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.731313 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-config-data-custom\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.731350 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-combined-ca-bundle\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.731399 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-public-tls-certs\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.731518 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nh5l\" (UniqueName: \"kubernetes.io/projected/23ac5ef2-d307-4073-982f-afe777413b41-kube-api-access-8nh5l\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.731573 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-config-data\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.777773 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.838175 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-internal-tls-certs\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.838612 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-config-data-custom\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.838636 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-public-tls-certs\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.838662 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-combined-ca-bundle\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.838721 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nh5l\" (UniqueName: \"kubernetes.io/projected/23ac5ef2-d307-4073-982f-afe777413b41-kube-api-access-8nh5l\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.838744 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-config-data\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.849555 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-public-tls-certs\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.850170 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-internal-tls-certs\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.851092 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-config-data-custom\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.852515 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-combined-ca-bundle\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.864713 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23ac5ef2-d307-4073-982f-afe777413b41-config-data\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.870147 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nh5l\" (UniqueName: \"kubernetes.io/projected/23ac5ef2-d307-4073-982f-afe777413b41-kube-api-access-8nh5l\") pod \"heat-api-56b4585b94-qjwmb\" (UID: \"23ac5ef2-d307-4073-982f-afe777413b41\") " pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:06 crc kubenswrapper[4912]: I1203 00:54:06.994186 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.159066 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.228505 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-6br9q"] Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.228898 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" podUID="c9e8199c-b459-4e55-88d6-f805d2ca430c" containerName="dnsmasq-dns" containerID="cri-o://f1dcbc6421bbb80fa82538457d9a82dfb6846533e0830ff975475ccb6c0132c6" gracePeriod=10 Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.307878 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-678985f55c-wftgz"] Dec 03 00:54:07 crc kubenswrapper[4912]: W1203 00:54:07.330203 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode31eb41a_2842_45fe_ba2d_6a84a048723f.slice/crio-f86454400ec99dbeb243ecb7d42ce7eed213ef04626386a49961fc1a885f9492 WatchSource:0}: Error finding container f86454400ec99dbeb243ecb7d42ce7eed213ef04626386a49961fc1a885f9492: Status 404 returned error can't find the container with id f86454400ec99dbeb243ecb7d42ce7eed213ef04626386a49961fc1a885f9492 Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.367683 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-678985f55c-wftgz" event={"ID":"e31eb41a-2842-45fe-ba2d-6a84a048723f","Type":"ContainerStarted","Data":"f86454400ec99dbeb243ecb7d42ce7eed213ef04626386a49961fc1a885f9492"} Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.371653 4912 generic.go:334] "Generic (PLEG): container finished" podID="c9e8199c-b459-4e55-88d6-f805d2ca430c" containerID="f1dcbc6421bbb80fa82538457d9a82dfb6846533e0830ff975475ccb6c0132c6" exitCode=0 Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.372708 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" event={"ID":"c9e8199c-b459-4e55-88d6-f805d2ca430c","Type":"ContainerDied","Data":"f1dcbc6421bbb80fa82538457d9a82dfb6846533e0830ff975475ccb6c0132c6"} Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.654946 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5cdb6985d5-xgpkb"] Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.671025 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-56b4585b94-qjwmb"] Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.847063 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.975860 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skh26\" (UniqueName: \"kubernetes.io/projected/c9e8199c-b459-4e55-88d6-f805d2ca430c-kube-api-access-skh26\") pod \"c9e8199c-b459-4e55-88d6-f805d2ca430c\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.975969 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-openstack-edpm-ipam\") pod \"c9e8199c-b459-4e55-88d6-f805d2ca430c\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.975990 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-svc\") pod \"c9e8199c-b459-4e55-88d6-f805d2ca430c\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.976057 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-swift-storage-0\") pod \"c9e8199c-b459-4e55-88d6-f805d2ca430c\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.976507 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-config\") pod \"c9e8199c-b459-4e55-88d6-f805d2ca430c\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.976727 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-nb\") pod \"c9e8199c-b459-4e55-88d6-f805d2ca430c\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.976760 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-sb\") pod \"c9e8199c-b459-4e55-88d6-f805d2ca430c\" (UID: \"c9e8199c-b459-4e55-88d6-f805d2ca430c\") " Dec 03 00:54:07 crc kubenswrapper[4912]: I1203 00:54:07.989694 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e8199c-b459-4e55-88d6-f805d2ca430c-kube-api-access-skh26" (OuterVolumeSpecName: "kube-api-access-skh26") pod "c9e8199c-b459-4e55-88d6-f805d2ca430c" (UID: "c9e8199c-b459-4e55-88d6-f805d2ca430c"). InnerVolumeSpecName "kube-api-access-skh26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.056650 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-config" (OuterVolumeSpecName: "config") pod "c9e8199c-b459-4e55-88d6-f805d2ca430c" (UID: "c9e8199c-b459-4e55-88d6-f805d2ca430c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.079558 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skh26\" (UniqueName: \"kubernetes.io/projected/c9e8199c-b459-4e55-88d6-f805d2ca430c-kube-api-access-skh26\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.079602 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-config\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.082974 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c9e8199c-b459-4e55-88d6-f805d2ca430c" (UID: "c9e8199c-b459-4e55-88d6-f805d2ca430c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.091222 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9e8199c-b459-4e55-88d6-f805d2ca430c" (UID: "c9e8199c-b459-4e55-88d6-f805d2ca430c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.098155 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9e8199c-b459-4e55-88d6-f805d2ca430c" (UID: "c9e8199c-b459-4e55-88d6-f805d2ca430c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.115872 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9e8199c-b459-4e55-88d6-f805d2ca430c" (UID: "c9e8199c-b459-4e55-88d6-f805d2ca430c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.128318 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "c9e8199c-b459-4e55-88d6-f805d2ca430c" (UID: "c9e8199c-b459-4e55-88d6-f805d2ca430c"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.186689 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.186742 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.186758 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.186776 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.186792 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9e8199c-b459-4e55-88d6-f805d2ca430c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.427832 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" event={"ID":"b0e666e2-5db8-409a-9a47-0da8fe895f75","Type":"ContainerStarted","Data":"18d79750e174f358df05ccfe3f414db9b67e988e45653db116112556b45b731a"} Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.432672 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-56b4585b94-qjwmb" event={"ID":"23ac5ef2-d307-4073-982f-afe777413b41","Type":"ContainerStarted","Data":"5982001ff306b74d10cce0140823b54617a9593d110c8363caab9678fa548456"} Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.437226 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-678985f55c-wftgz" event={"ID":"e31eb41a-2842-45fe-ba2d-6a84a048723f","Type":"ContainerStarted","Data":"0e0170f07483eefa155bb3ac64af72e5faf14f7f3a4507160085abb183409411"} Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.437923 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.453679 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" event={"ID":"c9e8199c-b459-4e55-88d6-f805d2ca430c","Type":"ContainerDied","Data":"7b8a0ace24aea71dbd58eec17da16779fdd76be85e9c589a3d3539a0e23004c2"} Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.453755 4912 scope.go:117] "RemoveContainer" containerID="f1dcbc6421bbb80fa82538457d9a82dfb6846533e0830ff975475ccb6c0132c6" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.454167 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-6br9q" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.467938 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-678985f55c-wftgz" podStartSLOduration=2.467841518 podStartE2EDuration="2.467841518s" podCreationTimestamp="2025-12-03 00:54:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:54:08.455531536 +0000 UTC m=+1834.097552106" watchObservedRunningTime="2025-12-03 00:54:08.467841518 +0000 UTC m=+1834.109862078" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.501012 4912 scope.go:117] "RemoveContainer" containerID="0e214b63380071b87afa976450b243427e8eda3f11328ae3f142246e965903dd" Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.509464 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-6br9q"] Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.521683 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-6br9q"] Dec 03 00:54:08 crc kubenswrapper[4912]: I1203 00:54:08.588693 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e8199c-b459-4e55-88d6-f805d2ca430c" path="/var/lib/kubelet/pods/c9e8199c-b459-4e55-88d6-f805d2ca430c/volumes" Dec 03 00:54:10 crc kubenswrapper[4912]: I1203 00:54:10.493033 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" event={"ID":"b0e666e2-5db8-409a-9a47-0da8fe895f75","Type":"ContainerStarted","Data":"fc88f1bf288a3101a93660e37cc1d4f2ad941a1505b4e0f0134db16fd45b635d"} Dec 03 00:54:10 crc kubenswrapper[4912]: I1203 00:54:10.494863 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:10 crc kubenswrapper[4912]: I1203 00:54:10.495924 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-56b4585b94-qjwmb" event={"ID":"23ac5ef2-d307-4073-982f-afe777413b41","Type":"ContainerStarted","Data":"85a46e864c8013fb6f31301faeadefdc89c94876294a311c30f311626f6b1d1a"} Dec 03 00:54:10 crc kubenswrapper[4912]: I1203 00:54:10.497616 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:10 crc kubenswrapper[4912]: I1203 00:54:10.524343 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" podStartSLOduration=2.398611842 podStartE2EDuration="4.524320366s" podCreationTimestamp="2025-12-03 00:54:06 +0000 UTC" firstStartedPulling="2025-12-03 00:54:07.665042576 +0000 UTC m=+1833.307063136" lastFinishedPulling="2025-12-03 00:54:09.7907511 +0000 UTC m=+1835.432771660" observedRunningTime="2025-12-03 00:54:10.516479144 +0000 UTC m=+1836.158499714" watchObservedRunningTime="2025-12-03 00:54:10.524320366 +0000 UTC m=+1836.166340926" Dec 03 00:54:10 crc kubenswrapper[4912]: I1203 00:54:10.539470 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-56b4585b94-qjwmb" podStartSLOduration=2.434360976 podStartE2EDuration="4.539448694s" podCreationTimestamp="2025-12-03 00:54:06 +0000 UTC" firstStartedPulling="2025-12-03 00:54:07.690071651 +0000 UTC m=+1833.332092211" lastFinishedPulling="2025-12-03 00:54:09.795159379 +0000 UTC m=+1835.437179929" observedRunningTime="2025-12-03 00:54:10.53556964 +0000 UTC m=+1836.177590230" watchObservedRunningTime="2025-12-03 00:54:10.539448694 +0000 UTC m=+1836.181469254" Dec 03 00:54:15 crc kubenswrapper[4912]: I1203 00:54:15.572561 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:54:15 crc kubenswrapper[4912]: E1203 00:54:15.574285 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.043764 4912 scope.go:117] "RemoveContainer" containerID="234e7fbb01e07f0821d49521459c233ab76b2224b69a548795bfcd2da6441e0b" Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.108513 4912 scope.go:117] "RemoveContainer" containerID="f7965bcbd8e1727a628460325f94bc010e3017cf31a0bad748076a1b407a4947" Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.141135 4912 scope.go:117] "RemoveContainer" containerID="e4a58ad548af6a4c89df80dca47420d9345f616bc171344b3e6777ff47a6e6eb" Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.215222 4912 scope.go:117] "RemoveContainer" containerID="02b42b411f8721761c9b9e86ca9993014177dc7bbcf2bcb5225128c8ff217eaf" Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.255060 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5cdb6985d5-xgpkb" Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.377389 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7b79dbdd58-qznp2"] Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.377663 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" podUID="d628e1f8-fd25-4bba-a85b-58f71eb64f58" containerName="heat-cfnapi" containerID="cri-o://bfd01aa91c6d8b8d3eec8616ba5eccf009db870e962ac170efa084202daab34d" gracePeriod=60 Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.425092 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-56b4585b94-qjwmb" Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.496649 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69f9575dc5-s25p4"] Dec 03 00:54:18 crc kubenswrapper[4912]: I1203 00:54:18.496893 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-69f9575dc5-s25p4" podUID="c7638b15-f910-4a20-b770-c65ebdc8ac49" containerName="heat-api" containerID="cri-o://c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488" gracePeriod=60 Dec 03 00:54:21 crc kubenswrapper[4912]: I1203 00:54:21.539446 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" podUID="d628e1f8-fd25-4bba-a85b-58f71eb64f58" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.209:8000/healthcheck\": read tcp 10.217.0.2:48756->10.217.0.209:8000: read: connection reset by peer" Dec 03 00:54:21 crc kubenswrapper[4912]: I1203 00:54:21.695067 4912 generic.go:334] "Generic (PLEG): container finished" podID="d628e1f8-fd25-4bba-a85b-58f71eb64f58" containerID="bfd01aa91c6d8b8d3eec8616ba5eccf009db870e962ac170efa084202daab34d" exitCode=0 Dec 03 00:54:21 crc kubenswrapper[4912]: I1203 00:54:21.695388 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" event={"ID":"d628e1f8-fd25-4bba-a85b-58f71eb64f58","Type":"ContainerDied","Data":"bfd01aa91c6d8b8d3eec8616ba5eccf009db870e962ac170efa084202daab34d"} Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.089919 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.173454 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.285020 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-internal-tls-certs\") pod \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.286368 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqmwj\" (UniqueName: \"kubernetes.io/projected/d628e1f8-fd25-4bba-a85b-58f71eb64f58-kube-api-access-dqmwj\") pod \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.286464 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data\") pod \"c7638b15-f910-4a20-b770-c65ebdc8ac49\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.286656 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-public-tls-certs\") pod \"c7638b15-f910-4a20-b770-c65ebdc8ac49\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.286705 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-internal-tls-certs\") pod \"c7638b15-f910-4a20-b770-c65ebdc8ac49\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.286735 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-public-tls-certs\") pod \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.286775 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data\") pod \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.286802 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q64n7\" (UniqueName: \"kubernetes.io/projected/c7638b15-f910-4a20-b770-c65ebdc8ac49-kube-api-access-q64n7\") pod \"c7638b15-f910-4a20-b770-c65ebdc8ac49\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.286827 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-combined-ca-bundle\") pod \"c7638b15-f910-4a20-b770-c65ebdc8ac49\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.287117 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-combined-ca-bundle\") pod \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.287187 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data-custom\") pod \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\" (UID: \"d628e1f8-fd25-4bba-a85b-58f71eb64f58\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.287221 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data-custom\") pod \"c7638b15-f910-4a20-b770-c65ebdc8ac49\" (UID: \"c7638b15-f910-4a20-b770-c65ebdc8ac49\") " Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.313680 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d628e1f8-fd25-4bba-a85b-58f71eb64f58-kube-api-access-dqmwj" (OuterVolumeSpecName: "kube-api-access-dqmwj") pod "d628e1f8-fd25-4bba-a85b-58f71eb64f58" (UID: "d628e1f8-fd25-4bba-a85b-58f71eb64f58"). InnerVolumeSpecName "kube-api-access-dqmwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.314747 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7638b15-f910-4a20-b770-c65ebdc8ac49-kube-api-access-q64n7" (OuterVolumeSpecName: "kube-api-access-q64n7") pod "c7638b15-f910-4a20-b770-c65ebdc8ac49" (UID: "c7638b15-f910-4a20-b770-c65ebdc8ac49"). InnerVolumeSpecName "kube-api-access-q64n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.320538 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c7638b15-f910-4a20-b770-c65ebdc8ac49" (UID: "c7638b15-f910-4a20-b770-c65ebdc8ac49"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.356221 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d628e1f8-fd25-4bba-a85b-58f71eb64f58" (UID: "d628e1f8-fd25-4bba-a85b-58f71eb64f58"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.384252 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7638b15-f910-4a20-b770-c65ebdc8ac49" (UID: "c7638b15-f910-4a20-b770-c65ebdc8ac49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.397037 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q64n7\" (UniqueName: \"kubernetes.io/projected/c7638b15-f910-4a20-b770-c65ebdc8ac49-kube-api-access-q64n7\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.397077 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.397088 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.397097 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.397108 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqmwj\" (UniqueName: \"kubernetes.io/projected/d628e1f8-fd25-4bba-a85b-58f71eb64f58-kube-api-access-dqmwj\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.412736 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c7638b15-f910-4a20-b770-c65ebdc8ac49" (UID: "c7638b15-f910-4a20-b770-c65ebdc8ac49"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.429038 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d628e1f8-fd25-4bba-a85b-58f71eb64f58" (UID: "d628e1f8-fd25-4bba-a85b-58f71eb64f58"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.429484 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data" (OuterVolumeSpecName: "config-data") pod "d628e1f8-fd25-4bba-a85b-58f71eb64f58" (UID: "d628e1f8-fd25-4bba-a85b-58f71eb64f58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.439503 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d628e1f8-fd25-4bba-a85b-58f71eb64f58" (UID: "d628e1f8-fd25-4bba-a85b-58f71eb64f58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.440619 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d628e1f8-fd25-4bba-a85b-58f71eb64f58" (UID: "d628e1f8-fd25-4bba-a85b-58f71eb64f58"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.453353 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data" (OuterVolumeSpecName: "config-data") pod "c7638b15-f910-4a20-b770-c65ebdc8ac49" (UID: "c7638b15-f910-4a20-b770-c65ebdc8ac49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.482706 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c7638b15-f910-4a20-b770-c65ebdc8ac49" (UID: "c7638b15-f910-4a20-b770-c65ebdc8ac49"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.499166 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.499204 4912 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.499223 4912 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7638b15-f910-4a20-b770-c65ebdc8ac49-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.499566 4912 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.499587 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.499610 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.499620 4912 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d628e1f8-fd25-4bba-a85b-58f71eb64f58-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.712531 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.713610 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7b79dbdd58-qznp2" event={"ID":"d628e1f8-fd25-4bba-a85b-58f71eb64f58","Type":"ContainerDied","Data":"1637ec03534bc02d4f4e665d6c348ab4bb4629ddba0c3722340a01da83c4e6d2"} Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.713668 4912 scope.go:117] "RemoveContainer" containerID="bfd01aa91c6d8b8d3eec8616ba5eccf009db870e962ac170efa084202daab34d" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.716060 4912 generic.go:334] "Generic (PLEG): container finished" podID="c7638b15-f910-4a20-b770-c65ebdc8ac49" containerID="c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488" exitCode=0 Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.716091 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69f9575dc5-s25p4" event={"ID":"c7638b15-f910-4a20-b770-c65ebdc8ac49","Type":"ContainerDied","Data":"c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488"} Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.716109 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69f9575dc5-s25p4" event={"ID":"c7638b15-f910-4a20-b770-c65ebdc8ac49","Type":"ContainerDied","Data":"c51d17de32b2f70c8fe305b5bad26b757164fd30cf9d372d7bfc24ae36731b81"} Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.716182 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69f9575dc5-s25p4" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.754471 4912 scope.go:117] "RemoveContainer" containerID="c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.771608 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7b79dbdd58-qznp2"] Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.803036 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7b79dbdd58-qznp2"] Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.814799 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69f9575dc5-s25p4"] Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.825610 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-69f9575dc5-s25p4"] Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.862654 4912 scope.go:117] "RemoveContainer" containerID="c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488" Dec 03 00:54:22 crc kubenswrapper[4912]: E1203 00:54:22.863278 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488\": container with ID starting with c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488 not found: ID does not exist" containerID="c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488" Dec 03 00:54:22 crc kubenswrapper[4912]: I1203 00:54:22.863326 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488"} err="failed to get container status \"c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488\": rpc error: code = NotFound desc = could not find container \"c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488\": container with ID starting with c7035da4317e833edd8d7dfaeb34f4217a3e32840b586c84b646398ba4af0488 not found: ID does not exist" Dec 03 00:54:24 crc kubenswrapper[4912]: I1203 00:54:24.617619 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7638b15-f910-4a20-b770-c65ebdc8ac49" path="/var/lib/kubelet/pods/c7638b15-f910-4a20-b770-c65ebdc8ac49/volumes" Dec 03 00:54:24 crc kubenswrapper[4912]: I1203 00:54:24.620474 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d628e1f8-fd25-4bba-a85b-58f71eb64f58" path="/var/lib/kubelet/pods/d628e1f8-fd25-4bba-a85b-58f71eb64f58/volumes" Dec 03 00:54:24 crc kubenswrapper[4912]: I1203 00:54:24.748655 4912 generic.go:334] "Generic (PLEG): container finished" podID="1435b2ef-93c1-41ac-a2bd-f9af839b7894" containerID="6aeda547d26dbbe49ce6f652c700f8f1325e7d5084c650583ca3df31f874f74a" exitCode=0 Dec 03 00:54:24 crc kubenswrapper[4912]: I1203 00:54:24.748762 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1435b2ef-93c1-41ac-a2bd-f9af839b7894","Type":"ContainerDied","Data":"6aeda547d26dbbe49ce6f652c700f8f1325e7d5084c650583ca3df31f874f74a"} Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.494242 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75"] Dec 03 00:54:25 crc kubenswrapper[4912]: E1203 00:54:25.494921 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d628e1f8-fd25-4bba-a85b-58f71eb64f58" containerName="heat-cfnapi" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.494942 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d628e1f8-fd25-4bba-a85b-58f71eb64f58" containerName="heat-cfnapi" Dec 03 00:54:25 crc kubenswrapper[4912]: E1203 00:54:25.494971 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e8199c-b459-4e55-88d6-f805d2ca430c" containerName="dnsmasq-dns" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.494980 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e8199c-b459-4e55-88d6-f805d2ca430c" containerName="dnsmasq-dns" Dec 03 00:54:25 crc kubenswrapper[4912]: E1203 00:54:25.495007 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e8199c-b459-4e55-88d6-f805d2ca430c" containerName="init" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.495017 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e8199c-b459-4e55-88d6-f805d2ca430c" containerName="init" Dec 03 00:54:25 crc kubenswrapper[4912]: E1203 00:54:25.495049 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7638b15-f910-4a20-b770-c65ebdc8ac49" containerName="heat-api" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.495058 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7638b15-f910-4a20-b770-c65ebdc8ac49" containerName="heat-api" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.495340 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d628e1f8-fd25-4bba-a85b-58f71eb64f58" containerName="heat-cfnapi" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.495374 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e8199c-b459-4e55-88d6-f805d2ca430c" containerName="dnsmasq-dns" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.495391 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7638b15-f910-4a20-b770-c65ebdc8ac49" containerName="heat-api" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.496496 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.508915 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.508997 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.508997 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.510557 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.527400 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75"] Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.625348 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.625622 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgsvb\" (UniqueName: \"kubernetes.io/projected/8e6b4897-744a-4650-b924-28a515fa875f-kube-api-access-jgsvb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.625738 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.625818 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.728361 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.728655 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.728718 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgsvb\" (UniqueName: \"kubernetes.io/projected/8e6b4897-744a-4650-b924-28a515fa875f-kube-api-access-jgsvb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.728784 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.736651 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.737402 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.752351 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.753955 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgsvb\" (UniqueName: \"kubernetes.io/projected/8e6b4897-744a-4650-b924-28a515fa875f-kube-api-access-jgsvb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-29v75\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.768463 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"1435b2ef-93c1-41ac-a2bd-f9af839b7894","Type":"ContainerStarted","Data":"47c39872371e8724e4fb30b105ecb0263635fc18147d22d3ec81c80d8ca0dd85"} Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.770513 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.773325 4912 generic.go:334] "Generic (PLEG): container finished" podID="eb7b3762-41c4-4935-8882-526576167e0a" containerID="5c5da28b70e3c9b4587c70196a56a8efa9e17074769861e73f05fa9fd2b0bfce" exitCode=0 Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.773407 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb7b3762-41c4-4935-8882-526576167e0a","Type":"ContainerDied","Data":"5c5da28b70e3c9b4587c70196a56a8efa9e17074769861e73f05fa9fd2b0bfce"} Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.803672 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=48.80358788 podStartE2EDuration="48.80358788s" podCreationTimestamp="2025-12-03 00:53:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:54:25.79284436 +0000 UTC m=+1851.434864950" watchObservedRunningTime="2025-12-03 00:54:25.80358788 +0000 UTC m=+1851.445608460" Dec 03 00:54:25 crc kubenswrapper[4912]: I1203 00:54:25.817845 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:26 crc kubenswrapper[4912]: W1203 00:54:26.450284 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e6b4897_744a_4650_b924_28a515fa875f.slice/crio-395b809226d618fbc83325fd2cf50bd210694fa3f41192b34b0f26f59d23bc0f WatchSource:0}: Error finding container 395b809226d618fbc83325fd2cf50bd210694fa3f41192b34b0f26f59d23bc0f: Status 404 returned error can't find the container with id 395b809226d618fbc83325fd2cf50bd210694fa3f41192b34b0f26f59d23bc0f Dec 03 00:54:26 crc kubenswrapper[4912]: I1203 00:54:26.455535 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75"] Dec 03 00:54:26 crc kubenswrapper[4912]: I1203 00:54:26.728169 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-678985f55c-wftgz" Dec 03 00:54:26 crc kubenswrapper[4912]: I1203 00:54:26.791519 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" event={"ID":"8e6b4897-744a-4650-b924-28a515fa875f","Type":"ContainerStarted","Data":"395b809226d618fbc83325fd2cf50bd210694fa3f41192b34b0f26f59d23bc0f"} Dec 03 00:54:26 crc kubenswrapper[4912]: I1203 00:54:26.796397 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"eb7b3762-41c4-4935-8882-526576167e0a","Type":"ContainerStarted","Data":"f684a48a722de5f7662cea70e4a5b02d93e27fd70dbcfc92448f30004e339f2a"} Dec 03 00:54:26 crc kubenswrapper[4912]: I1203 00:54:26.797230 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 00:54:26 crc kubenswrapper[4912]: I1203 00:54:26.805439 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5fc6c75d4d-bwk2j"] Dec 03 00:54:26 crc kubenswrapper[4912]: I1203 00:54:26.805689 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" podUID="74900d46-53f3-4cd5-aa56-9509c80b0041" containerName="heat-engine" containerID="cri-o://9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a" gracePeriod=60 Dec 03 00:54:26 crc kubenswrapper[4912]: I1203 00:54:26.855698 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=49.855668327 podStartE2EDuration="49.855668327s" podCreationTimestamp="2025-12-03 00:53:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 00:54:26.83833908 +0000 UTC m=+1852.480359650" watchObservedRunningTime="2025-12-03 00:54:26.855668327 +0000 UTC m=+1852.497688917" Dec 03 00:54:27 crc kubenswrapper[4912]: I1203 00:54:27.571862 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:54:27 crc kubenswrapper[4912]: E1203 00:54:27.572185 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:54:30 crc kubenswrapper[4912]: E1203 00:54:30.165847 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 00:54:30 crc kubenswrapper[4912]: E1203 00:54:30.167863 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 00:54:30 crc kubenswrapper[4912]: E1203 00:54:30.169304 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 00:54:30 crc kubenswrapper[4912]: E1203 00:54:30.169339 4912 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" podUID="74900d46-53f3-4cd5-aa56-9509c80b0041" containerName="heat-engine" Dec 03 00:54:35 crc kubenswrapper[4912]: I1203 00:54:35.856508 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-z6bqt"] Dec 03 00:54:35 crc kubenswrapper[4912]: I1203 00:54:35.877174 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-z6bqt"] Dec 03 00:54:35 crc kubenswrapper[4912]: I1203 00:54:35.942500 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-lfbmg"] Dec 03 00:54:35 crc kubenswrapper[4912]: I1203 00:54:35.944833 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:35 crc kubenswrapper[4912]: I1203 00:54:35.949670 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 00:54:35 crc kubenswrapper[4912]: I1203 00:54:35.993132 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-lfbmg"] Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.105987 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-combined-ca-bundle\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.106050 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l72j\" (UniqueName: \"kubernetes.io/projected/1de6def0-35f9-411e-9d55-475a18686b2c-kube-api-access-5l72j\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.106102 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-scripts\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.106177 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-config-data\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.212174 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-config-data\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.212355 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-combined-ca-bundle\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.212453 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l72j\" (UniqueName: \"kubernetes.io/projected/1de6def0-35f9-411e-9d55-475a18686b2c-kube-api-access-5l72j\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.212500 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-scripts\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.228922 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-combined-ca-bundle\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.232529 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-config-data\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.233770 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-scripts\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.236456 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l72j\" (UniqueName: \"kubernetes.io/projected/1de6def0-35f9-411e-9d55-475a18686b2c-kube-api-access-5l72j\") pod \"aodh-db-sync-lfbmg\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.306668 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:36 crc kubenswrapper[4912]: I1203 00:54:36.591970 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfd2dfcd-2539-425b-9071-d0de6bc9a598" path="/var/lib/kubelet/pods/cfd2dfcd-2539-425b-9071-d0de6bc9a598/volumes" Dec 03 00:54:38 crc kubenswrapper[4912]: I1203 00:54:38.224129 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 00:54:38 crc kubenswrapper[4912]: I1203 00:54:38.285943 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 00:54:38 crc kubenswrapper[4912]: I1203 00:54:38.928928 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 00:54:39 crc kubenswrapper[4912]: I1203 00:54:39.357019 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-lfbmg"] Dec 03 00:54:39 crc kubenswrapper[4912]: W1203 00:54:39.363968 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1de6def0_35f9_411e_9d55_475a18686b2c.slice/crio-de2f6ec019f78cd51f405e6df470e2587628a8f4771cb25d06905696ad4ccb3f WatchSource:0}: Error finding container de2f6ec019f78cd51f405e6df470e2587628a8f4771cb25d06905696ad4ccb3f: Status 404 returned error can't find the container with id de2f6ec019f78cd51f405e6df470e2587628a8f4771cb25d06905696ad4ccb3f Dec 03 00:54:40 crc kubenswrapper[4912]: I1203 00:54:40.137053 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-lfbmg" event={"ID":"1de6def0-35f9-411e-9d55-475a18686b2c","Type":"ContainerStarted","Data":"de2f6ec019f78cd51f405e6df470e2587628a8f4771cb25d06905696ad4ccb3f"} Dec 03 00:54:40 crc kubenswrapper[4912]: I1203 00:54:40.140098 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" event={"ID":"8e6b4897-744a-4650-b924-28a515fa875f","Type":"ContainerStarted","Data":"d89f010efa30b12367cce5e47a57d3a2aa8b796588c87fc9f5bc4d3ab46c953a"} Dec 03 00:54:40 crc kubenswrapper[4912]: E1203 00:54:40.166422 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 00:54:40 crc kubenswrapper[4912]: I1203 00:54:40.170852 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" podStartSLOduration=2.7022717309999997 podStartE2EDuration="15.170828436s" podCreationTimestamp="2025-12-03 00:54:25 +0000 UTC" firstStartedPulling="2025-12-03 00:54:26.452930564 +0000 UTC m=+1852.094951134" lastFinishedPulling="2025-12-03 00:54:38.921487279 +0000 UTC m=+1864.563507839" observedRunningTime="2025-12-03 00:54:40.15840838 +0000 UTC m=+1865.800428940" watchObservedRunningTime="2025-12-03 00:54:40.170828436 +0000 UTC m=+1865.812848996" Dec 03 00:54:40 crc kubenswrapper[4912]: E1203 00:54:40.176679 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 00:54:40 crc kubenswrapper[4912]: E1203 00:54:40.178484 4912 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 00:54:40 crc kubenswrapper[4912]: E1203 00:54:40.178522 4912 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" podUID="74900d46-53f3-4cd5-aa56-9509c80b0041" containerName="heat-engine" Dec 03 00:54:41 crc kubenswrapper[4912]: I1203 00:54:41.572835 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:54:41 crc kubenswrapper[4912]: E1203 00:54:41.573939 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 00:54:43 crc kubenswrapper[4912]: I1203 00:54:43.309739 4912 generic.go:334] "Generic (PLEG): container finished" podID="74900d46-53f3-4cd5-aa56-9509c80b0041" containerID="9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a" exitCode=0 Dec 03 00:54:43 crc kubenswrapper[4912]: I1203 00:54:43.309799 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" event={"ID":"74900d46-53f3-4cd5-aa56-9509c80b0041","Type":"ContainerDied","Data":"9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a"} Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.328086 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-lfbmg" event={"ID":"1de6def0-35f9-411e-9d55-475a18686b2c","Type":"ContainerStarted","Data":"04e8915b53cbbc4ff07d2af13a228d43ca0690a32e050c805a1efe4c43f6540f"} Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.331778 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" event={"ID":"74900d46-53f3-4cd5-aa56-9509c80b0041","Type":"ContainerDied","Data":"c51c2285879b2a463630686ee09f4c9518c07cf4e564c88f69a394ee817729f0"} Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.331849 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c51c2285879b2a463630686ee09f4c9518c07cf4e564c88f69a394ee817729f0" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.360872 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.378873 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data\") pod \"74900d46-53f3-4cd5-aa56-9509c80b0041\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.379146 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data-custom\") pod \"74900d46-53f3-4cd5-aa56-9509c80b0041\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.379251 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxzqm\" (UniqueName: \"kubernetes.io/projected/74900d46-53f3-4cd5-aa56-9509c80b0041-kube-api-access-jxzqm\") pod \"74900d46-53f3-4cd5-aa56-9509c80b0041\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.379365 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-combined-ca-bundle\") pod \"74900d46-53f3-4cd5-aa56-9509c80b0041\" (UID: \"74900d46-53f3-4cd5-aa56-9509c80b0041\") " Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.380759 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-lfbmg" podStartSLOduration=4.7862385530000005 podStartE2EDuration="9.380736056s" podCreationTimestamp="2025-12-03 00:54:35 +0000 UTC" firstStartedPulling="2025-12-03 00:54:39.367127658 +0000 UTC m=+1865.009148218" lastFinishedPulling="2025-12-03 00:54:43.961625161 +0000 UTC m=+1869.603645721" observedRunningTime="2025-12-03 00:54:44.364958669 +0000 UTC m=+1870.006979239" watchObservedRunningTime="2025-12-03 00:54:44.380736056 +0000 UTC m=+1870.022756616" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.386553 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "74900d46-53f3-4cd5-aa56-9509c80b0041" (UID: "74900d46-53f3-4cd5-aa56-9509c80b0041"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.394482 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74900d46-53f3-4cd5-aa56-9509c80b0041-kube-api-access-jxzqm" (OuterVolumeSpecName: "kube-api-access-jxzqm") pod "74900d46-53f3-4cd5-aa56-9509c80b0041" (UID: "74900d46-53f3-4cd5-aa56-9509c80b0041"). InnerVolumeSpecName "kube-api-access-jxzqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.434763 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74900d46-53f3-4cd5-aa56-9509c80b0041" (UID: "74900d46-53f3-4cd5-aa56-9509c80b0041"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.476152 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data" (OuterVolumeSpecName: "config-data") pod "74900d46-53f3-4cd5-aa56-9509c80b0041" (UID: "74900d46-53f3-4cd5-aa56-9509c80b0041"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.481986 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.482024 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.482039 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxzqm\" (UniqueName: \"kubernetes.io/projected/74900d46-53f3-4cd5-aa56-9509c80b0041-kube-api-access-jxzqm\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:44 crc kubenswrapper[4912]: I1203 00:54:44.482049 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74900d46-53f3-4cd5-aa56-9509c80b0041-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:45 crc kubenswrapper[4912]: I1203 00:54:45.342010 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5fc6c75d4d-bwk2j" Dec 03 00:54:45 crc kubenswrapper[4912]: I1203 00:54:45.379539 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5fc6c75d4d-bwk2j"] Dec 03 00:54:45 crc kubenswrapper[4912]: I1203 00:54:45.393049 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-5fc6c75d4d-bwk2j"] Dec 03 00:54:46 crc kubenswrapper[4912]: I1203 00:54:46.588465 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74900d46-53f3-4cd5-aa56-9509c80b0041" path="/var/lib/kubelet/pods/74900d46-53f3-4cd5-aa56-9509c80b0041/volumes" Dec 03 00:54:47 crc kubenswrapper[4912]: I1203 00:54:47.369150 4912 generic.go:334] "Generic (PLEG): container finished" podID="1de6def0-35f9-411e-9d55-475a18686b2c" containerID="04e8915b53cbbc4ff07d2af13a228d43ca0690a32e050c805a1efe4c43f6540f" exitCode=0 Dec 03 00:54:47 crc kubenswrapper[4912]: I1203 00:54:47.369298 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-lfbmg" event={"ID":"1de6def0-35f9-411e-9d55-475a18686b2c","Type":"ContainerDied","Data":"04e8915b53cbbc4ff07d2af13a228d43ca0690a32e050c805a1efe4c43f6540f"} Dec 03 00:54:48 crc kubenswrapper[4912]: I1203 00:54:48.878658 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.021784 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-combined-ca-bundle\") pod \"1de6def0-35f9-411e-9d55-475a18686b2c\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.021917 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l72j\" (UniqueName: \"kubernetes.io/projected/1de6def0-35f9-411e-9d55-475a18686b2c-kube-api-access-5l72j\") pod \"1de6def0-35f9-411e-9d55-475a18686b2c\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.022093 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-scripts\") pod \"1de6def0-35f9-411e-9d55-475a18686b2c\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.022258 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-config-data\") pod \"1de6def0-35f9-411e-9d55-475a18686b2c\" (UID: \"1de6def0-35f9-411e-9d55-475a18686b2c\") " Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.029834 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de6def0-35f9-411e-9d55-475a18686b2c-kube-api-access-5l72j" (OuterVolumeSpecName: "kube-api-access-5l72j") pod "1de6def0-35f9-411e-9d55-475a18686b2c" (UID: "1de6def0-35f9-411e-9d55-475a18686b2c"). InnerVolumeSpecName "kube-api-access-5l72j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.031117 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-scripts" (OuterVolumeSpecName: "scripts") pod "1de6def0-35f9-411e-9d55-475a18686b2c" (UID: "1de6def0-35f9-411e-9d55-475a18686b2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.067649 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1de6def0-35f9-411e-9d55-475a18686b2c" (UID: "1de6def0-35f9-411e-9d55-475a18686b2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.075588 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-config-data" (OuterVolumeSpecName: "config-data") pod "1de6def0-35f9-411e-9d55-475a18686b2c" (UID: "1de6def0-35f9-411e-9d55-475a18686b2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.125311 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.125369 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.125388 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l72j\" (UniqueName: \"kubernetes.io/projected/1de6def0-35f9-411e-9d55-475a18686b2c-kube-api-access-5l72j\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.125403 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1de6def0-35f9-411e-9d55-475a18686b2c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.400603 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-lfbmg" event={"ID":"1de6def0-35f9-411e-9d55-475a18686b2c","Type":"ContainerDied","Data":"de2f6ec019f78cd51f405e6df470e2587628a8f4771cb25d06905696ad4ccb3f"} Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.400662 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de2f6ec019f78cd51f405e6df470e2587628a8f4771cb25d06905696ad4ccb3f" Dec 03 00:54:49 crc kubenswrapper[4912]: I1203 00:54:49.400770 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-lfbmg" Dec 03 00:54:51 crc kubenswrapper[4912]: I1203 00:54:51.032965 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 00:54:51 crc kubenswrapper[4912]: I1203 00:54:51.431055 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-api" containerID="cri-o://391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5" gracePeriod=30 Dec 03 00:54:51 crc kubenswrapper[4912]: I1203 00:54:51.431112 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-listener" containerID="cri-o://f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd" gracePeriod=30 Dec 03 00:54:51 crc kubenswrapper[4912]: I1203 00:54:51.431220 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-evaluator" containerID="cri-o://710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75" gracePeriod=30 Dec 03 00:54:51 crc kubenswrapper[4912]: I1203 00:54:51.431207 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-notifier" containerID="cri-o://2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a" gracePeriod=30 Dec 03 00:54:52 crc kubenswrapper[4912]: I1203 00:54:52.452315 4912 generic.go:334] "Generic (PLEG): container finished" podID="8e6b4897-744a-4650-b924-28a515fa875f" containerID="d89f010efa30b12367cce5e47a57d3a2aa8b796588c87fc9f5bc4d3ab46c953a" exitCode=0 Dec 03 00:54:52 crc kubenswrapper[4912]: I1203 00:54:52.452393 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" event={"ID":"8e6b4897-744a-4650-b924-28a515fa875f","Type":"ContainerDied","Data":"d89f010efa30b12367cce5e47a57d3a2aa8b796588c87fc9f5bc4d3ab46c953a"} Dec 03 00:54:52 crc kubenswrapper[4912]: I1203 00:54:52.457467 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerDied","Data":"710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75"} Dec 03 00:54:52 crc kubenswrapper[4912]: I1203 00:54:52.457471 4912 generic.go:334] "Generic (PLEG): container finished" podID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerID="710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75" exitCode=0 Dec 03 00:54:52 crc kubenswrapper[4912]: I1203 00:54:52.457541 4912 generic.go:334] "Generic (PLEG): container finished" podID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerID="391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5" exitCode=0 Dec 03 00:54:52 crc kubenswrapper[4912]: I1203 00:54:52.457567 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerDied","Data":"391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5"} Dec 03 00:54:53 crc kubenswrapper[4912]: I1203 00:54:53.573179 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.106000 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.296222 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-ssh-key\") pod \"8e6b4897-744a-4650-b924-28a515fa875f\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.296478 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgsvb\" (UniqueName: \"kubernetes.io/projected/8e6b4897-744a-4650-b924-28a515fa875f-kube-api-access-jgsvb\") pod \"8e6b4897-744a-4650-b924-28a515fa875f\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.296681 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-repo-setup-combined-ca-bundle\") pod \"8e6b4897-744a-4650-b924-28a515fa875f\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.296754 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-inventory\") pod \"8e6b4897-744a-4650-b924-28a515fa875f\" (UID: \"8e6b4897-744a-4650-b924-28a515fa875f\") " Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.306106 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "8e6b4897-744a-4650-b924-28a515fa875f" (UID: "8e6b4897-744a-4650-b924-28a515fa875f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.306350 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6b4897-744a-4650-b924-28a515fa875f-kube-api-access-jgsvb" (OuterVolumeSpecName: "kube-api-access-jgsvb") pod "8e6b4897-744a-4650-b924-28a515fa875f" (UID: "8e6b4897-744a-4650-b924-28a515fa875f"). InnerVolumeSpecName "kube-api-access-jgsvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.332130 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8e6b4897-744a-4650-b924-28a515fa875f" (UID: "8e6b4897-744a-4650-b924-28a515fa875f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.335772 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-inventory" (OuterVolumeSpecName: "inventory") pod "8e6b4897-744a-4650-b924-28a515fa875f" (UID: "8e6b4897-744a-4650-b924-28a515fa875f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.399829 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.399885 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgsvb\" (UniqueName: \"kubernetes.io/projected/8e6b4897-744a-4650-b924-28a515fa875f-kube-api-access-jgsvb\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.399945 4912 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.399969 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b4897-744a-4650-b924-28a515fa875f-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.490645 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" event={"ID":"8e6b4897-744a-4650-b924-28a515fa875f","Type":"ContainerDied","Data":"395b809226d618fbc83325fd2cf50bd210694fa3f41192b34b0f26f59d23bc0f"} Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.491151 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="395b809226d618fbc83325fd2cf50bd210694fa3f41192b34b0f26f59d23bc0f" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.490919 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.505274 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"87c39cfe128fba833e7df8ba7ab177c6dc006107e5cd523a76fe1db299c9b196"} Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.612153 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx"] Dec 03 00:54:54 crc kubenswrapper[4912]: E1203 00:54:54.612790 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74900d46-53f3-4cd5-aa56-9509c80b0041" containerName="heat-engine" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.612807 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="74900d46-53f3-4cd5-aa56-9509c80b0041" containerName="heat-engine" Dec 03 00:54:54 crc kubenswrapper[4912]: E1203 00:54:54.612821 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de6def0-35f9-411e-9d55-475a18686b2c" containerName="aodh-db-sync" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.612828 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de6def0-35f9-411e-9d55-475a18686b2c" containerName="aodh-db-sync" Dec 03 00:54:54 crc kubenswrapper[4912]: E1203 00:54:54.612848 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6b4897-744a-4650-b924-28a515fa875f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.612856 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6b4897-744a-4650-b924-28a515fa875f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.613083 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="74900d46-53f3-4cd5-aa56-9509c80b0041" containerName="heat-engine" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.613104 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6b4897-744a-4650-b924-28a515fa875f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.613125 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de6def0-35f9-411e-9d55-475a18686b2c" containerName="aodh-db-sync" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.614043 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.619940 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.620190 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.620411 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.620592 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.624987 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx"] Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.711256 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.711452 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.711501 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.711752 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl6xs\" (UniqueName: \"kubernetes.io/projected/1c3f1295-f4b6-449f-b20e-229f2e89a32a-kube-api-access-fl6xs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.815057 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.815164 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.815242 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl6xs\" (UniqueName: \"kubernetes.io/projected/1c3f1295-f4b6-449f-b20e-229f2e89a32a-kube-api-access-fl6xs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.815420 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.824682 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.824852 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.837380 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl6xs\" (UniqueName: \"kubernetes.io/projected/1c3f1295-f4b6-449f-b20e-229f2e89a32a-kube-api-access-fl6xs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.845575 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:54 crc kubenswrapper[4912]: I1203 00:54:54.939547 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:54:55 crc kubenswrapper[4912]: I1203 00:54:55.579479 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx"] Dec 03 00:54:56 crc kubenswrapper[4912]: I1203 00:54:56.532746 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" event={"ID":"1c3f1295-f4b6-449f-b20e-229f2e89a32a","Type":"ContainerStarted","Data":"76942f72ae2455ee656d3d27d0d69e7868590808d99ca5e616938969d6beb5c3"} Dec 03 00:54:56 crc kubenswrapper[4912]: I1203 00:54:56.533827 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" event={"ID":"1c3f1295-f4b6-449f-b20e-229f2e89a32a","Type":"ContainerStarted","Data":"afe5f89631d736c30f9c761a07d1956535f41c406bb3e9f1431b0a83feb340ba"} Dec 03 00:54:56 crc kubenswrapper[4912]: I1203 00:54:56.565731 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" podStartSLOduration=2.128826736 podStartE2EDuration="2.565704319s" podCreationTimestamp="2025-12-03 00:54:54 +0000 UTC" firstStartedPulling="2025-12-03 00:54:55.58469966 +0000 UTC m=+1881.226720250" lastFinishedPulling="2025-12-03 00:54:56.021577243 +0000 UTC m=+1881.663597833" observedRunningTime="2025-12-03 00:54:56.553587783 +0000 UTC m=+1882.195608353" watchObservedRunningTime="2025-12-03 00:54:56.565704319 +0000 UTC m=+1882.207724889" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.262725 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.394720 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-internal-tls-certs\") pod \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.395244 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-scripts\") pod \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.395277 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-combined-ca-bundle\") pod \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.395369 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-public-tls-certs\") pod \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.395720 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-config-data\") pod \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.396085 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl7mq\" (UniqueName: \"kubernetes.io/projected/ddca249a-5151-469a-b9c0-34c5ad0ecc86-kube-api-access-dl7mq\") pod \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\" (UID: \"ddca249a-5151-469a-b9c0-34c5ad0ecc86\") " Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.403879 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-scripts" (OuterVolumeSpecName: "scripts") pod "ddca249a-5151-469a-b9c0-34c5ad0ecc86" (UID: "ddca249a-5151-469a-b9c0-34c5ad0ecc86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.427979 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddca249a-5151-469a-b9c0-34c5ad0ecc86-kube-api-access-dl7mq" (OuterVolumeSpecName: "kube-api-access-dl7mq") pod "ddca249a-5151-469a-b9c0-34c5ad0ecc86" (UID: "ddca249a-5151-469a-b9c0-34c5ad0ecc86"). InnerVolumeSpecName "kube-api-access-dl7mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.491126 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ddca249a-5151-469a-b9c0-34c5ad0ecc86" (UID: "ddca249a-5151-469a-b9c0-34c5ad0ecc86"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.502343 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.502381 4912 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.502400 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl7mq\" (UniqueName: \"kubernetes.io/projected/ddca249a-5151-469a-b9c0-34c5ad0ecc86-kube-api-access-dl7mq\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.507219 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ddca249a-5151-469a-b9c0-34c5ad0ecc86" (UID: "ddca249a-5151-469a-b9c0-34c5ad0ecc86"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.570743 4912 generic.go:334] "Generic (PLEG): container finished" podID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerID="f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd" exitCode=0 Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.570796 4912 generic.go:334] "Generic (PLEG): container finished" podID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerID="2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a" exitCode=0 Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.571578 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.572530 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerDied","Data":"f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd"} Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.572574 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerDied","Data":"2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a"} Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.572588 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ddca249a-5151-469a-b9c0-34c5ad0ecc86","Type":"ContainerDied","Data":"95a69893430f83bb8d0ba6688777b1b0e90e8b49ceae110ae446e1cd0f0c58ab"} Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.572605 4912 scope.go:117] "RemoveContainer" containerID="f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.608519 4912 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.618852 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddca249a-5151-469a-b9c0-34c5ad0ecc86" (UID: "ddca249a-5151-469a-b9c0-34c5ad0ecc86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.625845 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-config-data" (OuterVolumeSpecName: "config-data") pod "ddca249a-5151-469a-b9c0-34c5ad0ecc86" (UID: "ddca249a-5151-469a-b9c0-34c5ad0ecc86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.638347 4912 scope.go:117] "RemoveContainer" containerID="2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.663713 4912 scope.go:117] "RemoveContainer" containerID="710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.700240 4912 scope.go:117] "RemoveContainer" containerID="391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.712418 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.713073 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddca249a-5151-469a-b9c0-34c5ad0ecc86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.742871 4912 scope.go:117] "RemoveContainer" containerID="f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd" Dec 03 00:54:57 crc kubenswrapper[4912]: E1203 00:54:57.743864 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd\": container with ID starting with f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd not found: ID does not exist" containerID="f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.743909 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd"} err="failed to get container status \"f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd\": rpc error: code = NotFound desc = could not find container \"f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd\": container with ID starting with f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd not found: ID does not exist" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.743950 4912 scope.go:117] "RemoveContainer" containerID="2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a" Dec 03 00:54:57 crc kubenswrapper[4912]: E1203 00:54:57.744363 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a\": container with ID starting with 2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a not found: ID does not exist" containerID="2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.744390 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a"} err="failed to get container status \"2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a\": rpc error: code = NotFound desc = could not find container \"2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a\": container with ID starting with 2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a not found: ID does not exist" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.744410 4912 scope.go:117] "RemoveContainer" containerID="710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75" Dec 03 00:54:57 crc kubenswrapper[4912]: E1203 00:54:57.744813 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75\": container with ID starting with 710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75 not found: ID does not exist" containerID="710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.744841 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75"} err="failed to get container status \"710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75\": rpc error: code = NotFound desc = could not find container \"710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75\": container with ID starting with 710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75 not found: ID does not exist" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.744861 4912 scope.go:117] "RemoveContainer" containerID="391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5" Dec 03 00:54:57 crc kubenswrapper[4912]: E1203 00:54:57.745379 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5\": container with ID starting with 391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5 not found: ID does not exist" containerID="391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.745413 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5"} err="failed to get container status \"391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5\": rpc error: code = NotFound desc = could not find container \"391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5\": container with ID starting with 391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5 not found: ID does not exist" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.745458 4912 scope.go:117] "RemoveContainer" containerID="f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.745959 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd"} err="failed to get container status \"f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd\": rpc error: code = NotFound desc = could not find container \"f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd\": container with ID starting with f1e106e55838ea69a93862eaa31219904605d337eb989a1018ccccd4cf1813cd not found: ID does not exist" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.745987 4912 scope.go:117] "RemoveContainer" containerID="2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.746409 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a"} err="failed to get container status \"2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a\": rpc error: code = NotFound desc = could not find container \"2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a\": container with ID starting with 2d6c59151dd4cba841b244c7fd34ccbfcb25968a6299458350f3ef6a84143d1a not found: ID does not exist" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.746454 4912 scope.go:117] "RemoveContainer" containerID="710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.746857 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75"} err="failed to get container status \"710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75\": rpc error: code = NotFound desc = could not find container \"710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75\": container with ID starting with 710136f4950918675f86b63d7e45d01cb23d595e2aee0ef5b89067f19d111f75 not found: ID does not exist" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.746889 4912 scope.go:117] "RemoveContainer" containerID="391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.747889 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5"} err="failed to get container status \"391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5\": rpc error: code = NotFound desc = could not find container \"391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5\": container with ID starting with 391028e4009b3a3eaeca6d8e1928756a62413ee7ab966a613dd3b920133592b5 not found: ID does not exist" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.925415 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.949692 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.964767 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 00:54:57 crc kubenswrapper[4912]: E1203 00:54:57.965381 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-api" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.965402 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-api" Dec 03 00:54:57 crc kubenswrapper[4912]: E1203 00:54:57.965443 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-notifier" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.965453 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-notifier" Dec 03 00:54:57 crc kubenswrapper[4912]: E1203 00:54:57.965483 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-listener" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.965490 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-listener" Dec 03 00:54:57 crc kubenswrapper[4912]: E1203 00:54:57.965515 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-evaluator" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.965520 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-evaluator" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.965725 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-api" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.965744 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-evaluator" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.965763 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-notifier" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.965774 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" containerName="aodh-listener" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.968038 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.971468 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.971468 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.971706 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tnsxp" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.971893 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.973659 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 03 00:54:57 crc kubenswrapper[4912]: I1203 00:54:57.995579 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.124001 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sz69\" (UniqueName: \"kubernetes.io/projected/41945ca7-6d5e-4875-b770-801813413434-kube-api-access-6sz69\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.124121 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-scripts\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.124217 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-public-tls-certs\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.124277 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-internal-tls-certs\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.124530 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-combined-ca-bundle\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.124688 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-config-data\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.227853 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-config-data\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.227958 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sz69\" (UniqueName: \"kubernetes.io/projected/41945ca7-6d5e-4875-b770-801813413434-kube-api-access-6sz69\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.228031 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-scripts\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.228113 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-public-tls-certs\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.228187 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-internal-tls-certs\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.228249 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-combined-ca-bundle\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.233907 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-combined-ca-bundle\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.237118 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-scripts\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.237709 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-config-data\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.241932 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-internal-tls-certs\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.245086 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41945ca7-6d5e-4875-b770-801813413434-public-tls-certs\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.250122 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sz69\" (UniqueName: \"kubernetes.io/projected/41945ca7-6d5e-4875-b770-801813413434-kube-api-access-6sz69\") pod \"aodh-0\" (UID: \"41945ca7-6d5e-4875-b770-801813413434\") " pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.289789 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.592786 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddca249a-5151-469a-b9c0-34c5ad0ecc86" path="/var/lib/kubelet/pods/ddca249a-5151-469a-b9c0-34c5ad0ecc86/volumes" Dec 03 00:54:58 crc kubenswrapper[4912]: I1203 00:54:58.842214 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 00:54:58 crc kubenswrapper[4912]: W1203 00:54:58.845166 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41945ca7_6d5e_4875_b770_801813413434.slice/crio-e518ab9461fb4096994a7ca670daa6d1eb5163e677d165dcd30683d57bf016f9 WatchSource:0}: Error finding container e518ab9461fb4096994a7ca670daa6d1eb5163e677d165dcd30683d57bf016f9: Status 404 returned error can't find the container with id e518ab9461fb4096994a7ca670daa6d1eb5163e677d165dcd30683d57bf016f9 Dec 03 00:54:59 crc kubenswrapper[4912]: I1203 00:54:59.607445 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"41945ca7-6d5e-4875-b770-801813413434","Type":"ContainerStarted","Data":"cbf8b35340998cedf1b0aba66015b76a07be7a027aeb7d1eb919e2d3b0d8c037"} Dec 03 00:54:59 crc kubenswrapper[4912]: I1203 00:54:59.608476 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"41945ca7-6d5e-4875-b770-801813413434","Type":"ContainerStarted","Data":"e518ab9461fb4096994a7ca670daa6d1eb5163e677d165dcd30683d57bf016f9"} Dec 03 00:55:01 crc kubenswrapper[4912]: I1203 00:55:01.717607 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"41945ca7-6d5e-4875-b770-801813413434","Type":"ContainerStarted","Data":"0f7e13ba0821d0aca28095a523fa56671d5a107237bdd642d5aa96e01779d150"} Dec 03 00:55:02 crc kubenswrapper[4912]: I1203 00:55:02.739991 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"41945ca7-6d5e-4875-b770-801813413434","Type":"ContainerStarted","Data":"8d9317084628c71a7988cce73b5c2c3032c20ff0da294714d549132795bc81ef"} Dec 03 00:55:03 crc kubenswrapper[4912]: I1203 00:55:03.779050 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"41945ca7-6d5e-4875-b770-801813413434","Type":"ContainerStarted","Data":"4b5a573417e87cd72e8fbfc6ee447d606f694b2c46bd046d911d4c7bddac051f"} Dec 03 00:55:03 crc kubenswrapper[4912]: I1203 00:55:03.815911 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.264097521 podStartE2EDuration="6.815891342s" podCreationTimestamp="2025-12-03 00:54:57 +0000 UTC" firstStartedPulling="2025-12-03 00:54:58.849157889 +0000 UTC m=+1884.491178469" lastFinishedPulling="2025-12-03 00:55:03.40095174 +0000 UTC m=+1889.042972290" observedRunningTime="2025-12-03 00:55:03.807893247 +0000 UTC m=+1889.449913807" watchObservedRunningTime="2025-12-03 00:55:03.815891342 +0000 UTC m=+1889.457911902" Dec 03 00:55:18 crc kubenswrapper[4912]: I1203 00:55:18.589871 4912 scope.go:117] "RemoveContainer" containerID="a8ae68d8961560dd4cb9207fc6ae82494366831638580b903327485478be6d64" Dec 03 00:55:18 crc kubenswrapper[4912]: I1203 00:55:18.637503 4912 scope.go:117] "RemoveContainer" containerID="4b6094bafd4156f577df81eff79bccccb845d5ef2b8db70f115914b199486a77" Dec 03 00:55:18 crc kubenswrapper[4912]: I1203 00:55:18.695099 4912 scope.go:117] "RemoveContainer" containerID="703d3f45acd79c76aa7eafa8dd2f66e5e31432ddc2f817a46e36cf439ad37dcd" Dec 03 00:55:18 crc kubenswrapper[4912]: I1203 00:55:18.758971 4912 scope.go:117] "RemoveContainer" containerID="7275f0f547fdd62b6c803cb398bd5a78fa8bd7ce443148ada8771928682cb990" Dec 03 00:55:18 crc kubenswrapper[4912]: I1203 00:55:18.802476 4912 scope.go:117] "RemoveContainer" containerID="f9de75e3c05c329990648266c6e6b659d67ed0ff31d1903c4e0fd280977eda9c" Dec 03 00:55:18 crc kubenswrapper[4912]: I1203 00:55:18.840776 4912 scope.go:117] "RemoveContainer" containerID="9f1699b7d36aca8216a0caf12a1d18f162ef66ab271b63e30b8ddea67b51299f" Dec 03 00:55:18 crc kubenswrapper[4912]: I1203 00:55:18.867774 4912 scope.go:117] "RemoveContainer" containerID="2a585e40cc6f95d08d378351d5a1ca2a190d4059b133042e0dbe2586ddfef79b" Dec 03 00:56:19 crc kubenswrapper[4912]: I1203 00:56:19.097912 4912 scope.go:117] "RemoveContainer" containerID="863faecd86ce70d02cb5598f0cac5a5c04f195b242803006c751c1daef68c275" Dec 03 00:56:19 crc kubenswrapper[4912]: I1203 00:56:19.131876 4912 scope.go:117] "RemoveContainer" containerID="9827f4bf1886bbad99a5f599602ed2a5368e497fc72d71a53db6aa8b42e3f64a" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.025083 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zgtf4"] Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.029544 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.061821 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zgtf4"] Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.166408 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-utilities\") pod \"redhat-operators-zgtf4\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.166500 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbkwh\" (UniqueName: \"kubernetes.io/projected/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-kube-api-access-sbkwh\") pod \"redhat-operators-zgtf4\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.166551 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-catalog-content\") pod \"redhat-operators-zgtf4\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.269329 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-catalog-content\") pod \"redhat-operators-zgtf4\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.269536 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-utilities\") pod \"redhat-operators-zgtf4\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.269586 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbkwh\" (UniqueName: \"kubernetes.io/projected/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-kube-api-access-sbkwh\") pod \"redhat-operators-zgtf4\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.269990 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-catalog-content\") pod \"redhat-operators-zgtf4\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.270392 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-utilities\") pod \"redhat-operators-zgtf4\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.292025 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbkwh\" (UniqueName: \"kubernetes.io/projected/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-kube-api-access-sbkwh\") pod \"redhat-operators-zgtf4\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.385369 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:56:52 crc kubenswrapper[4912]: I1203 00:56:52.985876 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zgtf4"] Dec 03 00:56:53 crc kubenswrapper[4912]: I1203 00:56:53.447017 4912 generic.go:334] "Generic (PLEG): container finished" podID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerID="9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9" exitCode=0 Dec 03 00:56:53 crc kubenswrapper[4912]: I1203 00:56:53.447111 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgtf4" event={"ID":"12e85b73-63de-4ba3-8790-04c9e5d3b1a0","Type":"ContainerDied","Data":"9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9"} Dec 03 00:56:53 crc kubenswrapper[4912]: I1203 00:56:53.447483 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgtf4" event={"ID":"12e85b73-63de-4ba3-8790-04c9e5d3b1a0","Type":"ContainerStarted","Data":"240c6d3eadce30466c893b299fa23e05361e9c23843a45f7ace3ad5277dd84c3"} Dec 03 00:56:54 crc kubenswrapper[4912]: I1203 00:56:54.478295 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgtf4" event={"ID":"12e85b73-63de-4ba3-8790-04c9e5d3b1a0","Type":"ContainerStarted","Data":"9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6"} Dec 03 00:56:58 crc kubenswrapper[4912]: I1203 00:56:58.529162 4912 generic.go:334] "Generic (PLEG): container finished" podID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerID="9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6" exitCode=0 Dec 03 00:56:58 crc kubenswrapper[4912]: I1203 00:56:58.529294 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgtf4" event={"ID":"12e85b73-63de-4ba3-8790-04c9e5d3b1a0","Type":"ContainerDied","Data":"9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6"} Dec 03 00:56:59 crc kubenswrapper[4912]: I1203 00:56:59.547844 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgtf4" event={"ID":"12e85b73-63de-4ba3-8790-04c9e5d3b1a0","Type":"ContainerStarted","Data":"1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40"} Dec 03 00:56:59 crc kubenswrapper[4912]: I1203 00:56:59.588604 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zgtf4" podStartSLOduration=3.058282465 podStartE2EDuration="8.588578966s" podCreationTimestamp="2025-12-03 00:56:51 +0000 UTC" firstStartedPulling="2025-12-03 00:56:53.451618585 +0000 UTC m=+1999.093639145" lastFinishedPulling="2025-12-03 00:56:58.981915056 +0000 UTC m=+2004.623935646" observedRunningTime="2025-12-03 00:56:59.573207153 +0000 UTC m=+2005.215227743" watchObservedRunningTime="2025-12-03 00:56:59.588578966 +0000 UTC m=+2005.230599536" Dec 03 00:57:02 crc kubenswrapper[4912]: I1203 00:57:02.385709 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:57:02 crc kubenswrapper[4912]: I1203 00:57:02.386234 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:57:03 crc kubenswrapper[4912]: I1203 00:57:03.444318 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zgtf4" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerName="registry-server" probeResult="failure" output=< Dec 03 00:57:03 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 00:57:03 crc kubenswrapper[4912]: > Dec 03 00:57:12 crc kubenswrapper[4912]: I1203 00:57:12.469546 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:57:12 crc kubenswrapper[4912]: I1203 00:57:12.542605 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:57:12 crc kubenswrapper[4912]: I1203 00:57:12.730087 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zgtf4"] Dec 03 00:57:13 crc kubenswrapper[4912]: I1203 00:57:13.768578 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zgtf4" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerName="registry-server" containerID="cri-o://1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40" gracePeriod=2 Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.377833 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.428564 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-utilities\") pod \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.428708 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbkwh\" (UniqueName: \"kubernetes.io/projected/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-kube-api-access-sbkwh\") pod \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.428935 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-catalog-content\") pod \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\" (UID: \"12e85b73-63de-4ba3-8790-04c9e5d3b1a0\") " Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.429942 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-utilities" (OuterVolumeSpecName: "utilities") pod "12e85b73-63de-4ba3-8790-04c9e5d3b1a0" (UID: "12e85b73-63de-4ba3-8790-04c9e5d3b1a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.441249 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-kube-api-access-sbkwh" (OuterVolumeSpecName: "kube-api-access-sbkwh") pod "12e85b73-63de-4ba3-8790-04c9e5d3b1a0" (UID: "12e85b73-63de-4ba3-8790-04c9e5d3b1a0"). InnerVolumeSpecName "kube-api-access-sbkwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.532951 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.533003 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbkwh\" (UniqueName: \"kubernetes.io/projected/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-kube-api-access-sbkwh\") on node \"crc\" DevicePath \"\"" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.577824 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12e85b73-63de-4ba3-8790-04c9e5d3b1a0" (UID: "12e85b73-63de-4ba3-8790-04c9e5d3b1a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.635927 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12e85b73-63de-4ba3-8790-04c9e5d3b1a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.788638 4912 generic.go:334] "Generic (PLEG): container finished" podID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerID="1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40" exitCode=0 Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.788689 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgtf4" event={"ID":"12e85b73-63de-4ba3-8790-04c9e5d3b1a0","Type":"ContainerDied","Data":"1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40"} Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.788723 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zgtf4" event={"ID":"12e85b73-63de-4ba3-8790-04c9e5d3b1a0","Type":"ContainerDied","Data":"240c6d3eadce30466c893b299fa23e05361e9c23843a45f7ace3ad5277dd84c3"} Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.788743 4912 scope.go:117] "RemoveContainer" containerID="1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.788791 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zgtf4" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.844456 4912 scope.go:117] "RemoveContainer" containerID="9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.849632 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zgtf4"] Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.865350 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zgtf4"] Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.882294 4912 scope.go:117] "RemoveContainer" containerID="9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.942069 4912 scope.go:117] "RemoveContainer" containerID="1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40" Dec 03 00:57:14 crc kubenswrapper[4912]: E1203 00:57:14.943765 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40\": container with ID starting with 1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40 not found: ID does not exist" containerID="1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.943834 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40"} err="failed to get container status \"1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40\": rpc error: code = NotFound desc = could not find container \"1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40\": container with ID starting with 1b730e190aa363126213f44fcfbf7638b91bdcd2ac38cdca8746dcb0fbc34b40 not found: ID does not exist" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.943899 4912 scope.go:117] "RemoveContainer" containerID="9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6" Dec 03 00:57:14 crc kubenswrapper[4912]: E1203 00:57:14.944791 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6\": container with ID starting with 9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6 not found: ID does not exist" containerID="9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.944827 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6"} err="failed to get container status \"9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6\": rpc error: code = NotFound desc = could not find container \"9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6\": container with ID starting with 9f2ed967476e46d79a44d00da3a4fe19a45a0fa74799265ccfda60239728b2d6 not found: ID does not exist" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.944855 4912 scope.go:117] "RemoveContainer" containerID="9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9" Dec 03 00:57:14 crc kubenswrapper[4912]: E1203 00:57:14.946267 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9\": container with ID starting with 9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9 not found: ID does not exist" containerID="9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9" Dec 03 00:57:14 crc kubenswrapper[4912]: I1203 00:57:14.946295 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9"} err="failed to get container status \"9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9\": rpc error: code = NotFound desc = could not find container \"9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9\": container with ID starting with 9beed712f42740a9d94c35564d2df679f18c7ec8fadefe5fa534132167bd32b9 not found: ID does not exist" Dec 03 00:57:16 crc kubenswrapper[4912]: I1203 00:57:16.592557 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" path="/var/lib/kubelet/pods/12e85b73-63de-4ba3-8790-04c9e5d3b1a0/volumes" Dec 03 00:57:18 crc kubenswrapper[4912]: I1203 00:57:18.078367 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:57:18 crc kubenswrapper[4912]: I1203 00:57:18.078744 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:57:19 crc kubenswrapper[4912]: I1203 00:57:19.247832 4912 scope.go:117] "RemoveContainer" containerID="32e1e4fd1793c50379ca55561fa1c88a2fbca765be6bdf90df800352be7290e5" Dec 03 00:57:19 crc kubenswrapper[4912]: I1203 00:57:19.294869 4912 scope.go:117] "RemoveContainer" containerID="4d8ef5d001aa2e826b865b2c60bc27c58ada9507f12132f98c5d6bbf394f8042" Dec 03 00:57:19 crc kubenswrapper[4912]: I1203 00:57:19.373367 4912 scope.go:117] "RemoveContainer" containerID="9b9e8f9697d2a56d7564f1fd10315324cd057bf556cf51049d076bf7644a6aa3" Dec 03 00:57:30 crc kubenswrapper[4912]: I1203 00:57:30.076336 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-8dng2"] Dec 03 00:57:30 crc kubenswrapper[4912]: I1203 00:57:30.098615 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-8dng2"] Dec 03 00:57:30 crc kubenswrapper[4912]: I1203 00:57:30.586398 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e4aa219-c983-4ed9-96e6-e6583d25d34d" path="/var/lib/kubelet/pods/8e4aa219-c983-4ed9-96e6-e6583d25d34d/volumes" Dec 03 00:57:33 crc kubenswrapper[4912]: I1203 00:57:33.082949 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-s88vh"] Dec 03 00:57:33 crc kubenswrapper[4912]: I1203 00:57:33.096703 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-1612-account-create-update-qg8wc"] Dec 03 00:57:33 crc kubenswrapper[4912]: I1203 00:57:33.107739 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-51b0-account-create-update-vqt9z"] Dec 03 00:57:33 crc kubenswrapper[4912]: I1203 00:57:33.118358 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-51b0-account-create-update-vqt9z"] Dec 03 00:57:33 crc kubenswrapper[4912]: I1203 00:57:33.129210 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-s88vh"] Dec 03 00:57:33 crc kubenswrapper[4912]: I1203 00:57:33.139713 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-1612-account-create-update-qg8wc"] Dec 03 00:57:34 crc kubenswrapper[4912]: I1203 00:57:34.590382 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a43da64-c71a-4ae0-af3a-bc0fb42538ee" path="/var/lib/kubelet/pods/5a43da64-c71a-4ae0-af3a-bc0fb42538ee/volumes" Dec 03 00:57:34 crc kubenswrapper[4912]: I1203 00:57:34.591890 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd" path="/var/lib/kubelet/pods/a2aa2f0a-d5e4-43e8-b868-560cc8c05bcd/volumes" Dec 03 00:57:34 crc kubenswrapper[4912]: I1203 00:57:34.592543 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec53e2fd-3f5c-4382-a815-5e92e3299249" path="/var/lib/kubelet/pods/ec53e2fd-3f5c-4382-a815-5e92e3299249/volumes" Dec 03 00:57:36 crc kubenswrapper[4912]: I1203 00:57:36.043900 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-91cb-account-create-update-7fmk8"] Dec 03 00:57:36 crc kubenswrapper[4912]: I1203 00:57:36.064245 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-91cb-account-create-update-7fmk8"] Dec 03 00:57:36 crc kubenswrapper[4912]: I1203 00:57:36.595043 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480a1412-80ec-4d41-bd91-ce994043438f" path="/var/lib/kubelet/pods/480a1412-80ec-4d41-bd91-ce994043438f/volumes" Dec 03 00:57:37 crc kubenswrapper[4912]: I1203 00:57:37.086825 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-fe58-account-create-update-l5dwm"] Dec 03 00:57:37 crc kubenswrapper[4912]: I1203 00:57:37.101499 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-htd7z"] Dec 03 00:57:37 crc kubenswrapper[4912]: I1203 00:57:37.114445 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-fe58-account-create-update-l5dwm"] Dec 03 00:57:37 crc kubenswrapper[4912]: I1203 00:57:37.131009 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-4dvfk"] Dec 03 00:57:37 crc kubenswrapper[4912]: I1203 00:57:37.145341 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-htd7z"] Dec 03 00:57:37 crc kubenswrapper[4912]: I1203 00:57:37.159990 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-4dvfk"] Dec 03 00:57:38 crc kubenswrapper[4912]: I1203 00:57:38.591936 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15829b7d-eed7-425e-bd19-c2e496495444" path="/var/lib/kubelet/pods/15829b7d-eed7-425e-bd19-c2e496495444/volumes" Dec 03 00:57:38 crc kubenswrapper[4912]: I1203 00:57:38.594132 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c8872b9-97a4-48d7-bf51-bab0b0a5cf03" path="/var/lib/kubelet/pods/3c8872b9-97a4-48d7-bf51-bab0b0a5cf03/volumes" Dec 03 00:57:38 crc kubenswrapper[4912]: I1203 00:57:38.595562 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeaa6d61-6f8d-44dd-bd9b-708c59ab6598" path="/var/lib/kubelet/pods/eeaa6d61-6f8d-44dd-bd9b-708c59ab6598/volumes" Dec 03 00:57:47 crc kubenswrapper[4912]: I1203 00:57:47.052392 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-rx69z"] Dec 03 00:57:47 crc kubenswrapper[4912]: I1203 00:57:47.063992 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-rx69z"] Dec 03 00:57:47 crc kubenswrapper[4912]: I1203 00:57:47.075305 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-8f42-account-create-update-2mnqm"] Dec 03 00:57:47 crc kubenswrapper[4912]: I1203 00:57:47.086680 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-8f42-account-create-update-2mnqm"] Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.077892 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.078380 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.351527 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-57mcq"] Dec 03 00:57:48 crc kubenswrapper[4912]: E1203 00:57:48.352198 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerName="extract-content" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.352212 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerName="extract-content" Dec 03 00:57:48 crc kubenswrapper[4912]: E1203 00:57:48.352235 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerName="extract-utilities" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.352243 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerName="extract-utilities" Dec 03 00:57:48 crc kubenswrapper[4912]: E1203 00:57:48.352274 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerName="registry-server" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.352281 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerName="registry-server" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.352611 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e85b73-63de-4ba3-8790-04c9e5d3b1a0" containerName="registry-server" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.354467 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.407517 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-57mcq"] Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.494312 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-catalog-content\") pod \"community-operators-57mcq\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.494379 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-utilities\") pod \"community-operators-57mcq\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.494413 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d964w\" (UniqueName: \"kubernetes.io/projected/e613f2c2-7f2e-42cd-bde1-36f7243efda6-kube-api-access-d964w\") pod \"community-operators-57mcq\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.584897 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2efd8e63-52ce-4ed9-baa0-c1a34b684766" path="/var/lib/kubelet/pods/2efd8e63-52ce-4ed9-baa0-c1a34b684766/volumes" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.586076 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7" path="/var/lib/kubelet/pods/f4157ec9-cf5e-4edc-b302-0fbb1d58c6b7/volumes" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.597755 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-catalog-content\") pod \"community-operators-57mcq\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.597807 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-utilities\") pod \"community-operators-57mcq\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.597847 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d964w\" (UniqueName: \"kubernetes.io/projected/e613f2c2-7f2e-42cd-bde1-36f7243efda6-kube-api-access-d964w\") pod \"community-operators-57mcq\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.598538 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-utilities\") pod \"community-operators-57mcq\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.598709 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-catalog-content\") pod \"community-operators-57mcq\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.627934 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d964w\" (UniqueName: \"kubernetes.io/projected/e613f2c2-7f2e-42cd-bde1-36f7243efda6-kube-api-access-d964w\") pod \"community-operators-57mcq\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:48 crc kubenswrapper[4912]: I1203 00:57:48.685876 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:49 crc kubenswrapper[4912]: I1203 00:57:49.210829 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-57mcq"] Dec 03 00:57:49 crc kubenswrapper[4912]: I1203 00:57:49.316517 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57mcq" event={"ID":"e613f2c2-7f2e-42cd-bde1-36f7243efda6","Type":"ContainerStarted","Data":"92f3f206503c508cd5ce4c3d39efd316a00a0b887a8a113f3b0f8258b7c11a45"} Dec 03 00:57:50 crc kubenswrapper[4912]: I1203 00:57:50.332915 4912 generic.go:334] "Generic (PLEG): container finished" podID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerID="fc47f98c63e78123d58b819634e0d1f2249d374c9c9eedd5ff2f3a7919b47a57" exitCode=0 Dec 03 00:57:50 crc kubenswrapper[4912]: I1203 00:57:50.334381 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57mcq" event={"ID":"e613f2c2-7f2e-42cd-bde1-36f7243efda6","Type":"ContainerDied","Data":"fc47f98c63e78123d58b819634e0d1f2249d374c9c9eedd5ff2f3a7919b47a57"} Dec 03 00:57:50 crc kubenswrapper[4912]: I1203 00:57:50.337817 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 00:57:52 crc kubenswrapper[4912]: I1203 00:57:52.366303 4912 generic.go:334] "Generic (PLEG): container finished" podID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerID="e9a00150216afbd49e7736dc7059e1aad530ecf2a5cef18766def1ffdce4ce64" exitCode=0 Dec 03 00:57:52 crc kubenswrapper[4912]: I1203 00:57:52.366398 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57mcq" event={"ID":"e613f2c2-7f2e-42cd-bde1-36f7243efda6","Type":"ContainerDied","Data":"e9a00150216afbd49e7736dc7059e1aad530ecf2a5cef18766def1ffdce4ce64"} Dec 03 00:57:53 crc kubenswrapper[4912]: I1203 00:57:53.386395 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57mcq" event={"ID":"e613f2c2-7f2e-42cd-bde1-36f7243efda6","Type":"ContainerStarted","Data":"4d52b5310ccca02aabc692064a274fb3bcb921f03c3862b175c51b7d555080ce"} Dec 03 00:57:53 crc kubenswrapper[4912]: I1203 00:57:53.413267 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-57mcq" podStartSLOduration=2.86741207 podStartE2EDuration="5.413239731s" podCreationTimestamp="2025-12-03 00:57:48 +0000 UTC" firstStartedPulling="2025-12-03 00:57:50.337516114 +0000 UTC m=+2055.979536674" lastFinishedPulling="2025-12-03 00:57:52.883343775 +0000 UTC m=+2058.525364335" observedRunningTime="2025-12-03 00:57:53.40849783 +0000 UTC m=+2059.050518430" watchObservedRunningTime="2025-12-03 00:57:53.413239731 +0000 UTC m=+2059.055260301" Dec 03 00:57:58 crc kubenswrapper[4912]: I1203 00:57:58.686088 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:58 crc kubenswrapper[4912]: I1203 00:57:58.686849 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:58 crc kubenswrapper[4912]: I1203 00:57:58.788023 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:59 crc kubenswrapper[4912]: I1203 00:57:59.542358 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:57:59 crc kubenswrapper[4912]: I1203 00:57:59.600640 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-57mcq"] Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.072875 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-rbqgg"] Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.087114 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-rbqgg"] Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.101168 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-4039-account-create-update-g9m7m"] Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.111262 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-852d-account-create-update-l7bkl"] Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.121034 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-htck5"] Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.130630 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-4039-account-create-update-g9m7m"] Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.139608 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-852d-account-create-update-l7bkl"] Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.148070 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-htck5"] Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.597944 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1edf7b4a-adf2-4e25-ac63-29900c38445a" path="/var/lib/kubelet/pods/1edf7b4a-adf2-4e25-ac63-29900c38445a/volumes" Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.600773 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30895cbf-b323-4348-962b-32e976dbed88" path="/var/lib/kubelet/pods/30895cbf-b323-4348-962b-32e976dbed88/volumes" Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.601906 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c0f80dc-12ee-4a46-bf09-eceed25e25d1" path="/var/lib/kubelet/pods/5c0f80dc-12ee-4a46-bf09-eceed25e25d1/volumes" Dec 03 00:58:00 crc kubenswrapper[4912]: I1203 00:58:00.604520 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6013018a-f205-4218-a8b2-42ea78b05aaa" path="/var/lib/kubelet/pods/6013018a-f205-4218-a8b2-42ea78b05aaa/volumes" Dec 03 00:58:01 crc kubenswrapper[4912]: I1203 00:58:01.052323 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5df8-account-create-update-jblhz"] Dec 03 00:58:01 crc kubenswrapper[4912]: I1203 00:58:01.070128 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-65a2-account-create-update-52ck4"] Dec 03 00:58:01 crc kubenswrapper[4912]: I1203 00:58:01.099890 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-shcjq"] Dec 03 00:58:01 crc kubenswrapper[4912]: I1203 00:58:01.117648 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-prdcm"] Dec 03 00:58:01 crc kubenswrapper[4912]: I1203 00:58:01.134881 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5df8-account-create-update-jblhz"] Dec 03 00:58:01 crc kubenswrapper[4912]: I1203 00:58:01.151456 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-65a2-account-create-update-52ck4"] Dec 03 00:58:01 crc kubenswrapper[4912]: I1203 00:58:01.164220 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-shcjq"] Dec 03 00:58:01 crc kubenswrapper[4912]: I1203 00:58:01.173694 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-prdcm"] Dec 03 00:58:01 crc kubenswrapper[4912]: I1203 00:58:01.490040 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-57mcq" podUID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerName="registry-server" containerID="cri-o://4d52b5310ccca02aabc692064a274fb3bcb921f03c3862b175c51b7d555080ce" gracePeriod=2 Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.507574 4912 generic.go:334] "Generic (PLEG): container finished" podID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerID="4d52b5310ccca02aabc692064a274fb3bcb921f03c3862b175c51b7d555080ce" exitCode=0 Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.508019 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57mcq" event={"ID":"e613f2c2-7f2e-42cd-bde1-36f7243efda6","Type":"ContainerDied","Data":"4d52b5310ccca02aabc692064a274fb3bcb921f03c3862b175c51b7d555080ce"} Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.508120 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-57mcq" event={"ID":"e613f2c2-7f2e-42cd-bde1-36f7243efda6","Type":"ContainerDied","Data":"92f3f206503c508cd5ce4c3d39efd316a00a0b887a8a113f3b0f8258b7c11a45"} Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.508165 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92f3f206503c508cd5ce4c3d39efd316a00a0b887a8a113f3b0f8258b7c11a45" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.593150 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00bd8d98-9354-44b8-a44d-76027a084f3c" path="/var/lib/kubelet/pods/00bd8d98-9354-44b8-a44d-76027a084f3c/volumes" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.594316 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2167d1d7-c169-47a1-ab99-6de7765dc790" path="/var/lib/kubelet/pods/2167d1d7-c169-47a1-ab99-6de7765dc790/volumes" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.595168 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53898bac-8281-4e12-8caf-c8149ed14a4c" path="/var/lib/kubelet/pods/53898bac-8281-4e12-8caf-c8149ed14a4c/volumes" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.596604 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7594e6bc-7dc5-44cc-8810-c1e972e71764" path="/var/lib/kubelet/pods/7594e6bc-7dc5-44cc-8810-c1e972e71764/volumes" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.614828 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.727229 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-utilities\") pod \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.728056 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-catalog-content\") pod \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.728147 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-utilities" (OuterVolumeSpecName: "utilities") pod "e613f2c2-7f2e-42cd-bde1-36f7243efda6" (UID: "e613f2c2-7f2e-42cd-bde1-36f7243efda6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.728181 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d964w\" (UniqueName: \"kubernetes.io/projected/e613f2c2-7f2e-42cd-bde1-36f7243efda6-kube-api-access-d964w\") pod \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\" (UID: \"e613f2c2-7f2e-42cd-bde1-36f7243efda6\") " Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.729316 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.735035 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e613f2c2-7f2e-42cd-bde1-36f7243efda6-kube-api-access-d964w" (OuterVolumeSpecName: "kube-api-access-d964w") pod "e613f2c2-7f2e-42cd-bde1-36f7243efda6" (UID: "e613f2c2-7f2e-42cd-bde1-36f7243efda6"). InnerVolumeSpecName "kube-api-access-d964w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.786390 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e613f2c2-7f2e-42cd-bde1-36f7243efda6" (UID: "e613f2c2-7f2e-42cd-bde1-36f7243efda6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.832885 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d964w\" (UniqueName: \"kubernetes.io/projected/e613f2c2-7f2e-42cd-bde1-36f7243efda6-kube-api-access-d964w\") on node \"crc\" DevicePath \"\"" Dec 03 00:58:02 crc kubenswrapper[4912]: I1203 00:58:02.832931 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e613f2c2-7f2e-42cd-bde1-36f7243efda6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 00:58:03 crc kubenswrapper[4912]: I1203 00:58:03.525394 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-57mcq" Dec 03 00:58:03 crc kubenswrapper[4912]: I1203 00:58:03.589634 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-57mcq"] Dec 03 00:58:03 crc kubenswrapper[4912]: I1203 00:58:03.605045 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-57mcq"] Dec 03 00:58:04 crc kubenswrapper[4912]: I1203 00:58:04.593280 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" path="/var/lib/kubelet/pods/e613f2c2-7f2e-42cd-bde1-36f7243efda6/volumes" Dec 03 00:58:15 crc kubenswrapper[4912]: I1203 00:58:15.052042 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-j97dl"] Dec 03 00:58:15 crc kubenswrapper[4912]: I1203 00:58:15.065049 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-j97dl"] Dec 03 00:58:16 crc kubenswrapper[4912]: I1203 00:58:16.604008 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="039cc29c-336b-4c17-9d7d-f08488d07478" path="/var/lib/kubelet/pods/039cc29c-336b-4c17-9d7d-f08488d07478/volumes" Dec 03 00:58:18 crc kubenswrapper[4912]: I1203 00:58:18.078500 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 00:58:18 crc kubenswrapper[4912]: I1203 00:58:18.078933 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 00:58:18 crc kubenswrapper[4912]: I1203 00:58:18.078997 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 00:58:18 crc kubenswrapper[4912]: I1203 00:58:18.080102 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"87c39cfe128fba833e7df8ba7ab177c6dc006107e5cd523a76fe1db299c9b196"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 00:58:18 crc kubenswrapper[4912]: I1203 00:58:18.080171 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://87c39cfe128fba833e7df8ba7ab177c6dc006107e5cd523a76fe1db299c9b196" gracePeriod=600 Dec 03 00:58:18 crc kubenswrapper[4912]: I1203 00:58:18.740521 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="87c39cfe128fba833e7df8ba7ab177c6dc006107e5cd523a76fe1db299c9b196" exitCode=0 Dec 03 00:58:18 crc kubenswrapper[4912]: I1203 00:58:18.740578 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"87c39cfe128fba833e7df8ba7ab177c6dc006107e5cd523a76fe1db299c9b196"} Dec 03 00:58:18 crc kubenswrapper[4912]: I1203 00:58:18.740971 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc"} Dec 03 00:58:18 crc kubenswrapper[4912]: I1203 00:58:18.741016 4912 scope.go:117] "RemoveContainer" containerID="fff99fe7d6c904273d51bb68fe05821ed2d9e36703b98ceb106b800ab3bde02a" Dec 03 00:58:19 crc kubenswrapper[4912]: I1203 00:58:19.545882 4912 scope.go:117] "RemoveContainer" containerID="94ab507a74139bebe0993d007af1bdb40614a1c74ee316763ce88acc4c210a9d" Dec 03 00:58:19 crc kubenswrapper[4912]: I1203 00:58:19.601046 4912 scope.go:117] "RemoveContainer" containerID="3048d2a3a377cdd828e0687fec238d6794925143e3639fb9306dbb3e8ef6cdde" Dec 03 00:58:19 crc kubenswrapper[4912]: I1203 00:58:19.667615 4912 scope.go:117] "RemoveContainer" containerID="858cd40b2fa60b78d79641c0e6072a9ed2ff2e74af9da2047bee6ac5253aa75f" Dec 03 00:58:19 crc kubenswrapper[4912]: I1203 00:58:19.711938 4912 scope.go:117] "RemoveContainer" containerID="65b597ed4bbc1350c58cf547afdde883dd0adf46b4211e515132b0e0ad6962b8" Dec 03 00:58:19 crc kubenswrapper[4912]: I1203 00:58:19.781011 4912 scope.go:117] "RemoveContainer" containerID="5b05c942a3905fb2a98c88eaaa3619c19235d97157d3c64423680cfe622581bf" Dec 03 00:58:19 crc kubenswrapper[4912]: I1203 00:58:19.828965 4912 scope.go:117] "RemoveContainer" containerID="dca309568638bb16bf1be681679992b85fff17331db10a4e2d5df45e0a0b1699" Dec 03 00:58:19 crc kubenswrapper[4912]: I1203 00:58:19.880529 4912 scope.go:117] "RemoveContainer" containerID="dacaa5226c583e45648c8127972fbbf6adbb1b2e6dcf39b886cd09669f3242aa" Dec 03 00:58:19 crc kubenswrapper[4912]: I1203 00:58:19.920237 4912 scope.go:117] "RemoveContainer" containerID="b89ae9f2492911940bbbafaa992074af07db0735db2f13da75c218796efcd454" Dec 03 00:58:19 crc kubenswrapper[4912]: I1203 00:58:19.996291 4912 scope.go:117] "RemoveContainer" containerID="a3e22e06b6778bb0eb35826c62cecb496b0bac44d109486edd2c15373a57a2ee" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.023340 4912 scope.go:117] "RemoveContainer" containerID="3c603903509e566fa9464d11c926c221ae776e6f127fc32229f3c6b5e8e535b6" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.045489 4912 scope.go:117] "RemoveContainer" containerID="097553effd16008a2285a5950f01f07946d1046482f3611ac7b810bfecfc9bd3" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.071794 4912 scope.go:117] "RemoveContainer" containerID="9f58bc4c1070589d8cd090a43fbacdacda88465c4e21f61cc174d5fafe050899" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.100397 4912 scope.go:117] "RemoveContainer" containerID="bc63152b5763f5c7b2480ea95d1884cf7987930c728804f066aea3f8e55ec7df" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.123617 4912 scope.go:117] "RemoveContainer" containerID="f27b4b82e446bd02a99dc89d09ea3966c4dd87335a069a2080d9eb3d4f20a933" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.154504 4912 scope.go:117] "RemoveContainer" containerID="86093197118c3d966d8793f772f8b048f69db99608dbd3a11d170f362a81911c" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.219008 4912 scope.go:117] "RemoveContainer" containerID="73649cd60fde4ac8bc10e0e8b451468eee12d3d2a23ab399e0c2a2461a725bd6" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.250701 4912 scope.go:117] "RemoveContainer" containerID="9f623134169317a8bf47fcb248283e4afcde2f88f3e6f1d9a232b152158ff021" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.277575 4912 scope.go:117] "RemoveContainer" containerID="dbf85cf9f3adc7e94fdaa3b6eab3f310c481536c24eaa07dd2f23abe0d163ff0" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.311069 4912 scope.go:117] "RemoveContainer" containerID="3e2b15f714846231ae7bcf82aafdc8d03d21ac586a603d17914dbabfe32f9cc3" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.343768 4912 scope.go:117] "RemoveContainer" containerID="3e822a25c7ff8f47757b89553f4fe0ee778f9cc74ed8f6cdf5a8df204da4f123" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.385051 4912 scope.go:117] "RemoveContainer" containerID="18732f729a350106351f369c1cb8ccc7f8a10f3bfc3f4f547aafa0b3d0486b83" Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.796370 4912 generic.go:334] "Generic (PLEG): container finished" podID="1c3f1295-f4b6-449f-b20e-229f2e89a32a" containerID="76942f72ae2455ee656d3d27d0d69e7868590808d99ca5e616938969d6beb5c3" exitCode=0 Dec 03 00:58:20 crc kubenswrapper[4912]: I1203 00:58:20.796535 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" event={"ID":"1c3f1295-f4b6-449f-b20e-229f2e89a32a","Type":"ContainerDied","Data":"76942f72ae2455ee656d3d27d0d69e7868590808d99ca5e616938969d6beb5c3"} Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.475081 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.506186 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-ssh-key\") pod \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.506417 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl6xs\" (UniqueName: \"kubernetes.io/projected/1c3f1295-f4b6-449f-b20e-229f2e89a32a-kube-api-access-fl6xs\") pod \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.506595 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-inventory\") pod \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.506685 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-bootstrap-combined-ca-bundle\") pod \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\" (UID: \"1c3f1295-f4b6-449f-b20e-229f2e89a32a\") " Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.522035 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1c3f1295-f4b6-449f-b20e-229f2e89a32a" (UID: "1c3f1295-f4b6-449f-b20e-229f2e89a32a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.532468 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c3f1295-f4b6-449f-b20e-229f2e89a32a-kube-api-access-fl6xs" (OuterVolumeSpecName: "kube-api-access-fl6xs") pod "1c3f1295-f4b6-449f-b20e-229f2e89a32a" (UID: "1c3f1295-f4b6-449f-b20e-229f2e89a32a"). InnerVolumeSpecName "kube-api-access-fl6xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.630956 4912 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.631015 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl6xs\" (UniqueName: \"kubernetes.io/projected/1c3f1295-f4b6-449f-b20e-229f2e89a32a-kube-api-access-fl6xs\") on node \"crc\" DevicePath \"\"" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.675062 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1c3f1295-f4b6-449f-b20e-229f2e89a32a" (UID: "1c3f1295-f4b6-449f-b20e-229f2e89a32a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.721593 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-inventory" (OuterVolumeSpecName: "inventory") pod "1c3f1295-f4b6-449f-b20e-229f2e89a32a" (UID: "1c3f1295-f4b6-449f-b20e-229f2e89a32a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.738066 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.738107 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c3f1295-f4b6-449f-b20e-229f2e89a32a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.829629 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" event={"ID":"1c3f1295-f4b6-449f-b20e-229f2e89a32a","Type":"ContainerDied","Data":"afe5f89631d736c30f9c761a07d1956535f41c406bb3e9f1431b0a83feb340ba"} Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.829692 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afe5f89631d736c30f9c761a07d1956535f41c406bb3e9f1431b0a83feb340ba" Dec 03 00:58:22 crc kubenswrapper[4912]: I1203 00:58:22.829736 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.660073 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp"] Dec 03 00:58:23 crc kubenswrapper[4912]: E1203 00:58:23.660987 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerName="extract-utilities" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.661012 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerName="extract-utilities" Dec 03 00:58:23 crc kubenswrapper[4912]: E1203 00:58:23.661050 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerName="extract-content" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.661061 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerName="extract-content" Dec 03 00:58:23 crc kubenswrapper[4912]: E1203 00:58:23.661098 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c3f1295-f4b6-449f-b20e-229f2e89a32a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.661112 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c3f1295-f4b6-449f-b20e-229f2e89a32a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 00:58:23 crc kubenswrapper[4912]: E1203 00:58:23.661135 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerName="registry-server" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.661145 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerName="registry-server" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.661531 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="e613f2c2-7f2e-42cd-bde1-36f7243efda6" containerName="registry-server" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.661919 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c3f1295-f4b6-449f-b20e-229f2e89a32a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.663225 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.668525 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.668813 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.669985 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.672143 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.674118 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp"] Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.870218 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5cngp\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.870785 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79w52\" (UniqueName: \"kubernetes.io/projected/78887e9b-6ce9-4fc7-92ec-3c0b26052704-kube-api-access-79w52\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5cngp\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.870885 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5cngp\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.973835 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79w52\" (UniqueName: \"kubernetes.io/projected/78887e9b-6ce9-4fc7-92ec-3c0b26052704-kube-api-access-79w52\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5cngp\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.973893 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5cngp\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.973984 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5cngp\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.980040 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5cngp\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:23 crc kubenswrapper[4912]: I1203 00:58:23.980070 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5cngp\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:24 crc kubenswrapper[4912]: I1203 00:58:24.002644 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79w52\" (UniqueName: \"kubernetes.io/projected/78887e9b-6ce9-4fc7-92ec-3c0b26052704-kube-api-access-79w52\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-5cngp\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:24 crc kubenswrapper[4912]: I1203 00:58:24.012703 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:58:24 crc kubenswrapper[4912]: I1203 00:58:24.714656 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp"] Dec 03 00:58:24 crc kubenswrapper[4912]: W1203 00:58:24.730746 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78887e9b_6ce9_4fc7_92ec_3c0b26052704.slice/crio-6fb22d17c70faaee680c16ac48777211cd9ad60ec49776147487f6de4c50be65 WatchSource:0}: Error finding container 6fb22d17c70faaee680c16ac48777211cd9ad60ec49776147487f6de4c50be65: Status 404 returned error can't find the container with id 6fb22d17c70faaee680c16ac48777211cd9ad60ec49776147487f6de4c50be65 Dec 03 00:58:24 crc kubenswrapper[4912]: I1203 00:58:24.859417 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" event={"ID":"78887e9b-6ce9-4fc7-92ec-3c0b26052704","Type":"ContainerStarted","Data":"6fb22d17c70faaee680c16ac48777211cd9ad60ec49776147487f6de4c50be65"} Dec 03 00:58:25 crc kubenswrapper[4912]: I1203 00:58:25.872725 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" event={"ID":"78887e9b-6ce9-4fc7-92ec-3c0b26052704","Type":"ContainerStarted","Data":"8906b82ec139ea8f3d6e0d906c1bf93b2cac357db6496ee4e9b181c9caea469b"} Dec 03 00:58:25 crc kubenswrapper[4912]: I1203 00:58:25.904140 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" podStartSLOduration=2.381025694 podStartE2EDuration="2.904116667s" podCreationTimestamp="2025-12-03 00:58:23 +0000 UTC" firstStartedPulling="2025-12-03 00:58:24.734412792 +0000 UTC m=+2090.376433362" lastFinishedPulling="2025-12-03 00:58:25.257503765 +0000 UTC m=+2090.899524335" observedRunningTime="2025-12-03 00:58:25.893133206 +0000 UTC m=+2091.535153796" watchObservedRunningTime="2025-12-03 00:58:25.904116667 +0000 UTC m=+2091.546137227" Dec 03 00:58:55 crc kubenswrapper[4912]: I1203 00:58:55.081277 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-vlh4p"] Dec 03 00:58:55 crc kubenswrapper[4912]: I1203 00:58:55.103389 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-tnx9g"] Dec 03 00:58:55 crc kubenswrapper[4912]: I1203 00:58:55.114815 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-vlh4p"] Dec 03 00:58:55 crc kubenswrapper[4912]: I1203 00:58:55.125878 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-tnx9g"] Dec 03 00:58:56 crc kubenswrapper[4912]: I1203 00:58:56.589627 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22a2b0ca-d062-446f-bcc5-6209916b6e89" path="/var/lib/kubelet/pods/22a2b0ca-d062-446f-bcc5-6209916b6e89/volumes" Dec 03 00:58:56 crc kubenswrapper[4912]: I1203 00:58:56.590852 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e760382f-ba1d-4d51-bdd0-2cd62315c4ae" path="/var/lib/kubelet/pods/e760382f-ba1d-4d51-bdd0-2cd62315c4ae/volumes" Dec 03 00:59:04 crc kubenswrapper[4912]: I1203 00:59:04.041377 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-wnqwp"] Dec 03 00:59:04 crc kubenswrapper[4912]: I1203 00:59:04.053999 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-wnqwp"] Dec 03 00:59:04 crc kubenswrapper[4912]: I1203 00:59:04.589727 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="879819c5-1e33-4e31-bdfb-89d74e5efe8b" path="/var/lib/kubelet/pods/879819c5-1e33-4e31-bdfb-89d74e5efe8b/volumes" Dec 03 00:59:09 crc kubenswrapper[4912]: I1203 00:59:09.095528 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-x5lrh"] Dec 03 00:59:09 crc kubenswrapper[4912]: I1203 00:59:09.110932 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-x5lrh"] Dec 03 00:59:09 crc kubenswrapper[4912]: I1203 00:59:09.122781 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dkrd4"] Dec 03 00:59:09 crc kubenswrapper[4912]: I1203 00:59:09.134655 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dkrd4"] Dec 03 00:59:10 crc kubenswrapper[4912]: I1203 00:59:10.596409 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4bd0e42-218d-47de-a16f-214d5f4d2910" path="/var/lib/kubelet/pods/d4bd0e42-218d-47de-a16f-214d5f4d2910/volumes" Dec 03 00:59:10 crc kubenswrapper[4912]: I1203 00:59:10.598094 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df5c7224-2a57-4931-98b4-78574fb2690b" path="/var/lib/kubelet/pods/df5c7224-2a57-4931-98b4-78574fb2690b/volumes" Dec 03 00:59:20 crc kubenswrapper[4912]: I1203 00:59:20.953165 4912 scope.go:117] "RemoveContainer" containerID="c8b3de5502e8daf8ff011cc14855b97c41f69269c1f76f6835a9e3ddba99d0d9" Dec 03 00:59:21 crc kubenswrapper[4912]: I1203 00:59:21.031043 4912 scope.go:117] "RemoveContainer" containerID="8a470aa50b1dbb46c7a395ef382ffd8d3cdf3b67b5b3eb030f1e84aa26b601e7" Dec 03 00:59:21 crc kubenswrapper[4912]: I1203 00:59:21.125389 4912 scope.go:117] "RemoveContainer" containerID="f85a5a5cf82b8a1257b409bbbd7e9414671a111408a649c0e4dec60ae935e4bb" Dec 03 00:59:21 crc kubenswrapper[4912]: I1203 00:59:21.175566 4912 scope.go:117] "RemoveContainer" containerID="10da5eb5c055fa52db79e62a49ca41e7146cd4d94cdcc498e5edc7b6ceb3a6c6" Dec 03 00:59:21 crc kubenswrapper[4912]: I1203 00:59:21.237053 4912 scope.go:117] "RemoveContainer" containerID="ac4118f51a831c59975ddac5999c44968efe58969edd7422dfc30604341e260d" Dec 03 00:59:24 crc kubenswrapper[4912]: I1203 00:59:24.085713 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-b5xlq"] Dec 03 00:59:24 crc kubenswrapper[4912]: I1203 00:59:24.095964 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-b5xlq"] Dec 03 00:59:24 crc kubenswrapper[4912]: I1203 00:59:24.606303 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3071752c-9584-4ac1-8746-1b3a4ad3f4a1" path="/var/lib/kubelet/pods/3071752c-9584-4ac1-8746-1b3a4ad3f4a1/volumes" Dec 03 00:59:45 crc kubenswrapper[4912]: I1203 00:59:45.334189 4912 generic.go:334] "Generic (PLEG): container finished" podID="78887e9b-6ce9-4fc7-92ec-3c0b26052704" containerID="8906b82ec139ea8f3d6e0d906c1bf93b2cac357db6496ee4e9b181c9caea469b" exitCode=0 Dec 03 00:59:45 crc kubenswrapper[4912]: I1203 00:59:45.334454 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" event={"ID":"78887e9b-6ce9-4fc7-92ec-3c0b26052704","Type":"ContainerDied","Data":"8906b82ec139ea8f3d6e0d906c1bf93b2cac357db6496ee4e9b181c9caea469b"} Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.038859 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.061140 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79w52\" (UniqueName: \"kubernetes.io/projected/78887e9b-6ce9-4fc7-92ec-3c0b26052704-kube-api-access-79w52\") pod \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.061268 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-inventory\") pod \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.061413 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-ssh-key\") pod \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\" (UID: \"78887e9b-6ce9-4fc7-92ec-3c0b26052704\") " Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.073689 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78887e9b-6ce9-4fc7-92ec-3c0b26052704-kube-api-access-79w52" (OuterVolumeSpecName: "kube-api-access-79w52") pod "78887e9b-6ce9-4fc7-92ec-3c0b26052704" (UID: "78887e9b-6ce9-4fc7-92ec-3c0b26052704"). InnerVolumeSpecName "kube-api-access-79w52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.102683 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-inventory" (OuterVolumeSpecName: "inventory") pod "78887e9b-6ce9-4fc7-92ec-3c0b26052704" (UID: "78887e9b-6ce9-4fc7-92ec-3c0b26052704"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.104155 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "78887e9b-6ce9-4fc7-92ec-3c0b26052704" (UID: "78887e9b-6ce9-4fc7-92ec-3c0b26052704"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.165381 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79w52\" (UniqueName: \"kubernetes.io/projected/78887e9b-6ce9-4fc7-92ec-3c0b26052704-kube-api-access-79w52\") on node \"crc\" DevicePath \"\"" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.165426 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.165459 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78887e9b-6ce9-4fc7-92ec-3c0b26052704-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.374058 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" event={"ID":"78887e9b-6ce9-4fc7-92ec-3c0b26052704","Type":"ContainerDied","Data":"6fb22d17c70faaee680c16ac48777211cd9ad60ec49776147487f6de4c50be65"} Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.374132 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fb22d17c70faaee680c16ac48777211cd9ad60ec49776147487f6de4c50be65" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.374180 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.523892 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg"] Dec 03 00:59:47 crc kubenswrapper[4912]: E1203 00:59:47.524733 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78887e9b-6ce9-4fc7-92ec-3c0b26052704" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.524774 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="78887e9b-6ce9-4fc7-92ec-3c0b26052704" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.525169 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="78887e9b-6ce9-4fc7-92ec-3c0b26052704" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.526562 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.529285 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.529406 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.531404 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.531462 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.543403 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg"] Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.576453 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.576546 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.576974 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n46q6\" (UniqueName: \"kubernetes.io/projected/85124524-226b-46f3-b4c1-3675bfc1c15f-kube-api-access-n46q6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.679664 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n46q6\" (UniqueName: \"kubernetes.io/projected/85124524-226b-46f3-b4c1-3675bfc1c15f-kube-api-access-n46q6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.679768 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.679888 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.685285 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.688070 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.714322 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n46q6\" (UniqueName: \"kubernetes.io/projected/85124524-226b-46f3-b4c1-3675bfc1c15f-kube-api-access-n46q6\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:47 crc kubenswrapper[4912]: I1203 00:59:47.896060 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:48 crc kubenswrapper[4912]: I1203 00:59:48.543963 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg"] Dec 03 00:59:49 crc kubenswrapper[4912]: I1203 00:59:49.406307 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" event={"ID":"85124524-226b-46f3-b4c1-3675bfc1c15f","Type":"ContainerStarted","Data":"b1ac1f75e00f07f6f097bf0b24de3ff3d53d3ea17916c7d0a7c7047cdfc2d665"} Dec 03 00:59:49 crc kubenswrapper[4912]: I1203 00:59:49.406824 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" event={"ID":"85124524-226b-46f3-b4c1-3675bfc1c15f","Type":"ContainerStarted","Data":"ff9ab4cc67e5783f0551732abb4916478fab16142dcc6d0f73f30292ad66f14f"} Dec 03 00:59:49 crc kubenswrapper[4912]: I1203 00:59:49.443750 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" podStartSLOduration=1.928723684 podStartE2EDuration="2.443714469s" podCreationTimestamp="2025-12-03 00:59:47 +0000 UTC" firstStartedPulling="2025-12-03 00:59:48.551171475 +0000 UTC m=+2174.193192065" lastFinishedPulling="2025-12-03 00:59:49.06616228 +0000 UTC m=+2174.708182850" observedRunningTime="2025-12-03 00:59:49.42967622 +0000 UTC m=+2175.071696800" watchObservedRunningTime="2025-12-03 00:59:49.443714469 +0000 UTC m=+2175.085735059" Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.062600 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-9697-account-create-update-chh6n"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.081525 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-fqvcx"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.103314 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-vk882"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.115584 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-9697-account-create-update-chh6n"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.128166 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-fqvcx"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.140813 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-24b5-account-create-update-4glzv"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.153585 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-81fe-account-create-update-8dcm2"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.169421 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-vk882"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.183062 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-4b7wk"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.203326 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-24b5-account-create-update-4glzv"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.214507 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-81fe-account-create-update-8dcm2"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.257957 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-4b7wk"] Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.591836 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d9ac0c8-42e7-4d85-be36-fd702a7c8787" path="/var/lib/kubelet/pods/3d9ac0c8-42e7-4d85-be36-fd702a7c8787/volumes" Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.593266 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9" path="/var/lib/kubelet/pods/4a828c9e-b47d-4166-ba1a-8a88d1d1f7d9/volumes" Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.593963 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cea1b95-ee14-4157-80f0-8785715d76fb" path="/var/lib/kubelet/pods/4cea1b95-ee14-4157-80f0-8785715d76fb/volumes" Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.594659 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e0fc17b-4a05-4eef-87d9-1269183df954" path="/var/lib/kubelet/pods/5e0fc17b-4a05-4eef-87d9-1269183df954/volumes" Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.596368 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f8ba108-0d7f-4399-a95f-aaf6b65b3499" path="/var/lib/kubelet/pods/8f8ba108-0d7f-4399-a95f-aaf6b65b3499/volumes" Dec 03 00:59:54 crc kubenswrapper[4912]: I1203 00:59:54.597128 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae1a6a50-b1c5-43b9-994c-7d3072dd8f72" path="/var/lib/kubelet/pods/ae1a6a50-b1c5-43b9-994c-7d3072dd8f72/volumes" Dec 03 00:59:55 crc kubenswrapper[4912]: I1203 00:59:55.475387 4912 generic.go:334] "Generic (PLEG): container finished" podID="85124524-226b-46f3-b4c1-3675bfc1c15f" containerID="b1ac1f75e00f07f6f097bf0b24de3ff3d53d3ea17916c7d0a7c7047cdfc2d665" exitCode=0 Dec 03 00:59:55 crc kubenswrapper[4912]: I1203 00:59:55.475484 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" event={"ID":"85124524-226b-46f3-b4c1-3675bfc1c15f","Type":"ContainerDied","Data":"b1ac1f75e00f07f6f097bf0b24de3ff3d53d3ea17916c7d0a7c7047cdfc2d665"} Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.064691 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.099230 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n46q6\" (UniqueName: \"kubernetes.io/projected/85124524-226b-46f3-b4c1-3675bfc1c15f-kube-api-access-n46q6\") pod \"85124524-226b-46f3-b4c1-3675bfc1c15f\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.099519 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-ssh-key\") pod \"85124524-226b-46f3-b4c1-3675bfc1c15f\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.099558 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-inventory\") pod \"85124524-226b-46f3-b4c1-3675bfc1c15f\" (UID: \"85124524-226b-46f3-b4c1-3675bfc1c15f\") " Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.108028 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85124524-226b-46f3-b4c1-3675bfc1c15f-kube-api-access-n46q6" (OuterVolumeSpecName: "kube-api-access-n46q6") pod "85124524-226b-46f3-b4c1-3675bfc1c15f" (UID: "85124524-226b-46f3-b4c1-3675bfc1c15f"). InnerVolumeSpecName "kube-api-access-n46q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.143088 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "85124524-226b-46f3-b4c1-3675bfc1c15f" (UID: "85124524-226b-46f3-b4c1-3675bfc1c15f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.166157 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-inventory" (OuterVolumeSpecName: "inventory") pod "85124524-226b-46f3-b4c1-3675bfc1c15f" (UID: "85124524-226b-46f3-b4c1-3675bfc1c15f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.205196 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n46q6\" (UniqueName: \"kubernetes.io/projected/85124524-226b-46f3-b4c1-3675bfc1c15f-kube-api-access-n46q6\") on node \"crc\" DevicePath \"\"" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.205248 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.205263 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/85124524-226b-46f3-b4c1-3675bfc1c15f-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.518183 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" event={"ID":"85124524-226b-46f3-b4c1-3675bfc1c15f","Type":"ContainerDied","Data":"ff9ab4cc67e5783f0551732abb4916478fab16142dcc6d0f73f30292ad66f14f"} Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.518242 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff9ab4cc67e5783f0551732abb4916478fab16142dcc6d0f73f30292ad66f14f" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.518334 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.736819 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg"] Dec 03 00:59:57 crc kubenswrapper[4912]: E1203 00:59:57.737917 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85124524-226b-46f3-b4c1-3675bfc1c15f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.737970 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="85124524-226b-46f3-b4c1-3675bfc1c15f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.738458 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="85124524-226b-46f3-b4c1-3675bfc1c15f" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.740049 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.746194 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.746425 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.747740 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.750028 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.760351 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg"] Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.820238 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svm4s\" (UniqueName: \"kubernetes.io/projected/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-kube-api-access-svm4s\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-stfcg\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.820461 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-stfcg\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.820636 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-stfcg\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.923916 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-stfcg\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.924121 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-stfcg\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.924264 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svm4s\" (UniqueName: \"kubernetes.io/projected/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-kube-api-access-svm4s\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-stfcg\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.932571 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-stfcg\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.933086 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-stfcg\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:57 crc kubenswrapper[4912]: I1203 00:59:57.955817 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svm4s\" (UniqueName: \"kubernetes.io/projected/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-kube-api-access-svm4s\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-stfcg\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:58 crc kubenswrapper[4912]: I1203 00:59:58.073135 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 00:59:58 crc kubenswrapper[4912]: I1203 00:59:58.739709 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg"] Dec 03 00:59:59 crc kubenswrapper[4912]: I1203 00:59:59.552132 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" event={"ID":"4b592c0a-8523-4584-9a2e-c616ea7d4ecd","Type":"ContainerStarted","Data":"de3211be03648379f08b63fa0bc862aad7ec27637a428906513f19421b1349c5"} Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.167259 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6"] Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.170390 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.173124 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.176820 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.183251 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6"] Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.330287 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f4130dc-bed8-4576-a343-2a41db43d477-config-volume\") pod \"collect-profiles-29412060-ktkz6\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.330461 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2nxg\" (UniqueName: \"kubernetes.io/projected/4f4130dc-bed8-4576-a343-2a41db43d477-kube-api-access-w2nxg\") pod \"collect-profiles-29412060-ktkz6\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.330838 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f4130dc-bed8-4576-a343-2a41db43d477-secret-volume\") pod \"collect-profiles-29412060-ktkz6\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.433473 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f4130dc-bed8-4576-a343-2a41db43d477-config-volume\") pod \"collect-profiles-29412060-ktkz6\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.433612 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2nxg\" (UniqueName: \"kubernetes.io/projected/4f4130dc-bed8-4576-a343-2a41db43d477-kube-api-access-w2nxg\") pod \"collect-profiles-29412060-ktkz6\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.433701 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f4130dc-bed8-4576-a343-2a41db43d477-secret-volume\") pod \"collect-profiles-29412060-ktkz6\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.434656 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f4130dc-bed8-4576-a343-2a41db43d477-config-volume\") pod \"collect-profiles-29412060-ktkz6\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.442046 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f4130dc-bed8-4576-a343-2a41db43d477-secret-volume\") pod \"collect-profiles-29412060-ktkz6\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.454583 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2nxg\" (UniqueName: \"kubernetes.io/projected/4f4130dc-bed8-4576-a343-2a41db43d477-kube-api-access-w2nxg\") pod \"collect-profiles-29412060-ktkz6\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.500498 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.604161 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" event={"ID":"4b592c0a-8523-4584-9a2e-c616ea7d4ecd","Type":"ContainerStarted","Data":"7aab6c5a1fa5d9a4cd1835efbff54d2dc216b26997bbe60f483abc7da0765283"} Dec 03 01:00:00 crc kubenswrapper[4912]: I1203 01:00:00.613576 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" podStartSLOduration=2.996251433 podStartE2EDuration="3.613544215s" podCreationTimestamp="2025-12-03 00:59:57 +0000 UTC" firstStartedPulling="2025-12-03 00:59:58.750333548 +0000 UTC m=+2184.392354108" lastFinishedPulling="2025-12-03 00:59:59.36762633 +0000 UTC m=+2185.009646890" observedRunningTime="2025-12-03 01:00:00.599100085 +0000 UTC m=+2186.241120675" watchObservedRunningTime="2025-12-03 01:00:00.613544215 +0000 UTC m=+2186.255564785" Dec 03 01:00:01 crc kubenswrapper[4912]: W1203 01:00:01.056559 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f4130dc_bed8_4576_a343_2a41db43d477.slice/crio-076635a18926c37ebd8517b3a91744f3d6e0622d27e448d3ebd884110e44cef6 WatchSource:0}: Error finding container 076635a18926c37ebd8517b3a91744f3d6e0622d27e448d3ebd884110e44cef6: Status 404 returned error can't find the container with id 076635a18926c37ebd8517b3a91744f3d6e0622d27e448d3ebd884110e44cef6 Dec 03 01:00:01 crc kubenswrapper[4912]: I1203 01:00:01.064005 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6"] Dec 03 01:00:01 crc kubenswrapper[4912]: I1203 01:00:01.590079 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" event={"ID":"4f4130dc-bed8-4576-a343-2a41db43d477","Type":"ContainerStarted","Data":"96a493340f606df04130503787a40dc53f344b6ebff9ef51e2c604722d3e532a"} Dec 03 01:00:01 crc kubenswrapper[4912]: I1203 01:00:01.590611 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" event={"ID":"4f4130dc-bed8-4576-a343-2a41db43d477","Type":"ContainerStarted","Data":"076635a18926c37ebd8517b3a91744f3d6e0622d27e448d3ebd884110e44cef6"} Dec 03 01:00:01 crc kubenswrapper[4912]: I1203 01:00:01.619972 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" podStartSLOduration=1.619946052 podStartE2EDuration="1.619946052s" podCreationTimestamp="2025-12-03 01:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:00:01.616656608 +0000 UTC m=+2187.258677178" watchObservedRunningTime="2025-12-03 01:00:01.619946052 +0000 UTC m=+2187.261966612" Dec 03 01:00:02 crc kubenswrapper[4912]: I1203 01:00:02.615664 4912 generic.go:334] "Generic (PLEG): container finished" podID="4f4130dc-bed8-4576-a343-2a41db43d477" containerID="96a493340f606df04130503787a40dc53f344b6ebff9ef51e2c604722d3e532a" exitCode=0 Dec 03 01:00:02 crc kubenswrapper[4912]: I1203 01:00:02.615746 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" event={"ID":"4f4130dc-bed8-4576-a343-2a41db43d477","Type":"ContainerDied","Data":"96a493340f606df04130503787a40dc53f344b6ebff9ef51e2c604722d3e532a"} Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.175069 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.265948 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f4130dc-bed8-4576-a343-2a41db43d477-secret-volume\") pod \"4f4130dc-bed8-4576-a343-2a41db43d477\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.266024 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2nxg\" (UniqueName: \"kubernetes.io/projected/4f4130dc-bed8-4576-a343-2a41db43d477-kube-api-access-w2nxg\") pod \"4f4130dc-bed8-4576-a343-2a41db43d477\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.266238 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f4130dc-bed8-4576-a343-2a41db43d477-config-volume\") pod \"4f4130dc-bed8-4576-a343-2a41db43d477\" (UID: \"4f4130dc-bed8-4576-a343-2a41db43d477\") " Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.266762 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f4130dc-bed8-4576-a343-2a41db43d477-config-volume" (OuterVolumeSpecName: "config-volume") pod "4f4130dc-bed8-4576-a343-2a41db43d477" (UID: "4f4130dc-bed8-4576-a343-2a41db43d477"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.285646 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f4130dc-bed8-4576-a343-2a41db43d477-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4f4130dc-bed8-4576-a343-2a41db43d477" (UID: "4f4130dc-bed8-4576-a343-2a41db43d477"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.289002 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f4130dc-bed8-4576-a343-2a41db43d477-kube-api-access-w2nxg" (OuterVolumeSpecName: "kube-api-access-w2nxg") pod "4f4130dc-bed8-4576-a343-2a41db43d477" (UID: "4f4130dc-bed8-4576-a343-2a41db43d477"). InnerVolumeSpecName "kube-api-access-w2nxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.369418 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4f4130dc-bed8-4576-a343-2a41db43d477-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.369725 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2nxg\" (UniqueName: \"kubernetes.io/projected/4f4130dc-bed8-4576-a343-2a41db43d477-kube-api-access-w2nxg\") on node \"crc\" DevicePath \"\"" Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.369805 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4f4130dc-bed8-4576-a343-2a41db43d477-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.668011 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" event={"ID":"4f4130dc-bed8-4576-a343-2a41db43d477","Type":"ContainerDied","Data":"076635a18926c37ebd8517b3a91744f3d6e0622d27e448d3ebd884110e44cef6"} Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.668078 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="076635a18926c37ebd8517b3a91744f3d6e0622d27e448d3ebd884110e44cef6" Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.668172 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6" Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.729905 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4"] Dec 03 01:00:04 crc kubenswrapper[4912]: I1203 01:00:04.739847 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412015-66lp4"] Dec 03 01:00:06 crc kubenswrapper[4912]: I1203 01:00:06.596273 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6a647db-07ba-49f6-9150-0a038d4fbee8" path="/var/lib/kubelet/pods/f6a647db-07ba-49f6-9150-0a038d4fbee8/volumes" Dec 03 01:00:18 crc kubenswrapper[4912]: I1203 01:00:18.078755 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:00:18 crc kubenswrapper[4912]: I1203 01:00:18.079647 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:00:21 crc kubenswrapper[4912]: I1203 01:00:21.440403 4912 scope.go:117] "RemoveContainer" containerID="9543c465d05ab8f926067fdf27ccce5b142898b3c761307e8731c643f36dcb34" Dec 03 01:00:21 crc kubenswrapper[4912]: I1203 01:00:21.489848 4912 scope.go:117] "RemoveContainer" containerID="41d342ee08eaa27f5b4fa2489afbec5b1bc97e532fc15e9f47b0efe8a1e3645f" Dec 03 01:00:21 crc kubenswrapper[4912]: I1203 01:00:21.538478 4912 scope.go:117] "RemoveContainer" containerID="be579efe3bbae50445d1e5c57ef8a64c44ba098f9355fcb4f7d3fee15461a1a9" Dec 03 01:00:21 crc kubenswrapper[4912]: I1203 01:00:21.589496 4912 scope.go:117] "RemoveContainer" containerID="6bc31f24cf6fb25378a406779e58e5e45a75472810588f9b4e613dadd172fdd5" Dec 03 01:00:21 crc kubenswrapper[4912]: I1203 01:00:21.646556 4912 scope.go:117] "RemoveContainer" containerID="f2326a066994c7938c4e11983325cefb03f174cff70bdfaf1cf803dbeca640eb" Dec 03 01:00:21 crc kubenswrapper[4912]: I1203 01:00:21.716090 4912 scope.go:117] "RemoveContainer" containerID="9396fe1c1097427efa3529e92e45ad6c7b7495250877f84663531e11bb6929d5" Dec 03 01:00:21 crc kubenswrapper[4912]: I1203 01:00:21.755718 4912 scope.go:117] "RemoveContainer" containerID="94a4977e9785851db0df0bd8592b4ba0d031198bade66107dccab490141f0daf" Dec 03 01:00:21 crc kubenswrapper[4912]: I1203 01:00:21.795037 4912 scope.go:117] "RemoveContainer" containerID="40ce820d269d557ca9832e26779b1fc816bf349512285458c543914afb6d317b" Dec 03 01:00:45 crc kubenswrapper[4912]: I1203 01:00:45.070839 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wpgr9"] Dec 03 01:00:45 crc kubenswrapper[4912]: I1203 01:00:45.083104 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-wpgr9"] Dec 03 01:00:45 crc kubenswrapper[4912]: I1203 01:00:45.223047 4912 generic.go:334] "Generic (PLEG): container finished" podID="4b592c0a-8523-4584-9a2e-c616ea7d4ecd" containerID="7aab6c5a1fa5d9a4cd1835efbff54d2dc216b26997bbe60f483abc7da0765283" exitCode=0 Dec 03 01:00:45 crc kubenswrapper[4912]: I1203 01:00:45.223100 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" event={"ID":"4b592c0a-8523-4584-9a2e-c616ea7d4ecd","Type":"ContainerDied","Data":"7aab6c5a1fa5d9a4cd1835efbff54d2dc216b26997bbe60f483abc7da0765283"} Dec 03 01:00:46 crc kubenswrapper[4912]: I1203 01:00:46.612048 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeac5989-b199-4fcf-8d5e-c49ac6062bfd" path="/var/lib/kubelet/pods/aeac5989-b199-4fcf-8d5e-c49ac6062bfd/volumes" Dec 03 01:00:46 crc kubenswrapper[4912]: I1203 01:00:46.894655 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.074066 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svm4s\" (UniqueName: \"kubernetes.io/projected/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-kube-api-access-svm4s\") pod \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.074300 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-ssh-key\") pod \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.074527 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-inventory\") pod \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\" (UID: \"4b592c0a-8523-4584-9a2e-c616ea7d4ecd\") " Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.102745 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-kube-api-access-svm4s" (OuterVolumeSpecName: "kube-api-access-svm4s") pod "4b592c0a-8523-4584-9a2e-c616ea7d4ecd" (UID: "4b592c0a-8523-4584-9a2e-c616ea7d4ecd"). InnerVolumeSpecName "kube-api-access-svm4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.120480 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-inventory" (OuterVolumeSpecName: "inventory") pod "4b592c0a-8523-4584-9a2e-c616ea7d4ecd" (UID: "4b592c0a-8523-4584-9a2e-c616ea7d4ecd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.121210 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b592c0a-8523-4584-9a2e-c616ea7d4ecd" (UID: "4b592c0a-8523-4584-9a2e-c616ea7d4ecd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.178198 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.178253 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.178274 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svm4s\" (UniqueName: \"kubernetes.io/projected/4b592c0a-8523-4584-9a2e-c616ea7d4ecd-kube-api-access-svm4s\") on node \"crc\" DevicePath \"\"" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.262710 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" event={"ID":"4b592c0a-8523-4584-9a2e-c616ea7d4ecd","Type":"ContainerDied","Data":"de3211be03648379f08b63fa0bc862aad7ec27637a428906513f19421b1349c5"} Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.262759 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de3211be03648379f08b63fa0bc862aad7ec27637a428906513f19421b1349c5" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.262910 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.429493 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx"] Dec 03 01:00:47 crc kubenswrapper[4912]: E1203 01:00:47.430177 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b592c0a-8523-4584-9a2e-c616ea7d4ecd" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.430196 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b592c0a-8523-4584-9a2e-c616ea7d4ecd" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:00:47 crc kubenswrapper[4912]: E1203 01:00:47.430215 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f4130dc-bed8-4576-a343-2a41db43d477" containerName="collect-profiles" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.430222 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f4130dc-bed8-4576-a343-2a41db43d477" containerName="collect-profiles" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.430459 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b592c0a-8523-4584-9a2e-c616ea7d4ecd" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.430481 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f4130dc-bed8-4576-a343-2a41db43d477" containerName="collect-profiles" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.431411 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.441016 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.441141 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.452176 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.461638 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.487534 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx"] Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.497532 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.497579 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.497634 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nchvs\" (UniqueName: \"kubernetes.io/projected/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-kube-api-access-nchvs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.600283 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.600636 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.600834 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nchvs\" (UniqueName: \"kubernetes.io/projected/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-kube-api-access-nchvs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.607018 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.612179 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.618123 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nchvs\" (UniqueName: \"kubernetes.io/projected/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-kube-api-access-nchvs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:47 crc kubenswrapper[4912]: I1203 01:00:47.797567 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:48 crc kubenswrapper[4912]: I1203 01:00:48.083926 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:00:48 crc kubenswrapper[4912]: I1203 01:00:48.084013 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:00:48 crc kubenswrapper[4912]: I1203 01:00:48.485996 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx"] Dec 03 01:00:49 crc kubenswrapper[4912]: I1203 01:00:49.292054 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" event={"ID":"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9","Type":"ContainerStarted","Data":"b06590253dfc4876443455e5e940e77f729a31da4dfefba8df2df9f2157d8216"} Dec 03 01:00:50 crc kubenswrapper[4912]: I1203 01:00:50.309765 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" event={"ID":"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9","Type":"ContainerStarted","Data":"04b81e41c9836381a227ffbc8251c0b5ca9855f82315cb9631ec842a805200d5"} Dec 03 01:00:50 crc kubenswrapper[4912]: I1203 01:00:50.335282 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" podStartSLOduration=2.535466736 podStartE2EDuration="3.335251859s" podCreationTimestamp="2025-12-03 01:00:47 +0000 UTC" firstStartedPulling="2025-12-03 01:00:48.480149567 +0000 UTC m=+2234.122170137" lastFinishedPulling="2025-12-03 01:00:49.27993469 +0000 UTC m=+2234.921955260" observedRunningTime="2025-12-03 01:00:50.327268975 +0000 UTC m=+2235.969289595" watchObservedRunningTime="2025-12-03 01:00:50.335251859 +0000 UTC m=+2235.977272409" Dec 03 01:00:54 crc kubenswrapper[4912]: I1203 01:00:54.366626 4912 generic.go:334] "Generic (PLEG): container finished" podID="c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9" containerID="04b81e41c9836381a227ffbc8251c0b5ca9855f82315cb9631ec842a805200d5" exitCode=0 Dec 03 01:00:54 crc kubenswrapper[4912]: I1203 01:00:54.366723 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" event={"ID":"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9","Type":"ContainerDied","Data":"04b81e41c9836381a227ffbc8251c0b5ca9855f82315cb9631ec842a805200d5"} Dec 03 01:00:55 crc kubenswrapper[4912]: I1203 01:00:55.872967 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.043884 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nchvs\" (UniqueName: \"kubernetes.io/projected/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-kube-api-access-nchvs\") pod \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.043999 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-inventory\") pod \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.044030 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-ssh-key\") pod \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\" (UID: \"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9\") " Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.053363 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-kube-api-access-nchvs" (OuterVolumeSpecName: "kube-api-access-nchvs") pod "c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9" (UID: "c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9"). InnerVolumeSpecName "kube-api-access-nchvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.091062 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-inventory" (OuterVolumeSpecName: "inventory") pod "c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9" (UID: "c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.100245 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9" (UID: "c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.148734 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nchvs\" (UniqueName: \"kubernetes.io/projected/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-kube-api-access-nchvs\") on node \"crc\" DevicePath \"\"" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.148841 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.148872 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.394163 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" event={"ID":"c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9","Type":"ContainerDied","Data":"b06590253dfc4876443455e5e940e77f729a31da4dfefba8df2df9f2157d8216"} Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.394238 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b06590253dfc4876443455e5e940e77f729a31da4dfefba8df2df9f2157d8216" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.394268 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.504294 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf"] Dec 03 01:00:56 crc kubenswrapper[4912]: E1203 01:00:56.505258 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.505383 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.506629 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.508178 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.510283 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.511129 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.511322 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.514872 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.518576 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf"] Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.571526 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfq8n\" (UniqueName: \"kubernetes.io/projected/0908ef60-cdec-46de-beb6-167295e04796-kube-api-access-jfq8n\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.571593 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.571631 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.672681 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfq8n\" (UniqueName: \"kubernetes.io/projected/0908ef60-cdec-46de-beb6-167295e04796-kube-api-access-jfq8n\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.673124 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.673168 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.681526 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.700484 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.705821 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfq8n\" (UniqueName: \"kubernetes.io/projected/0908ef60-cdec-46de-beb6-167295e04796-kube-api-access-jfq8n\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:56 crc kubenswrapper[4912]: I1203 01:00:56.840295 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:00:57 crc kubenswrapper[4912]: I1203 01:00:57.567822 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf"] Dec 03 01:00:58 crc kubenswrapper[4912]: I1203 01:00:58.444628 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" event={"ID":"0908ef60-cdec-46de-beb6-167295e04796","Type":"ContainerStarted","Data":"5c9d9eff1acf7d358d730edc5e20382faa3aca1799e3ace928e84fadea14bc80"} Dec 03 01:00:58 crc kubenswrapper[4912]: I1203 01:00:58.450639 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" event={"ID":"0908ef60-cdec-46de-beb6-167295e04796","Type":"ContainerStarted","Data":"90ba08b2e7f09f42cd7b2937cb60c5578c912972f6d65914424bf176d7d31bc2"} Dec 03 01:00:58 crc kubenswrapper[4912]: I1203 01:00:58.475904 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" podStartSLOduration=1.962179951 podStartE2EDuration="2.475884572s" podCreationTimestamp="2025-12-03 01:00:56 +0000 UTC" firstStartedPulling="2025-12-03 01:00:57.57707179 +0000 UTC m=+2243.219092350" lastFinishedPulling="2025-12-03 01:00:58.090776401 +0000 UTC m=+2243.732796971" observedRunningTime="2025-12-03 01:00:58.468623718 +0000 UTC m=+2244.110644278" watchObservedRunningTime="2025-12-03 01:00:58.475884572 +0000 UTC m=+2244.117905132" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.152522 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29412061-jvbbz"] Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.155213 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.178824 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412061-jvbbz"] Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.198826 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-combined-ca-bundle\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.199002 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqb2\" (UniqueName: \"kubernetes.io/projected/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-kube-api-access-nrqb2\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.199054 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-config-data\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.199329 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-fernet-keys\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.300063 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-config-data\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.300504 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-fernet-keys\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.300661 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-combined-ca-bundle\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.300781 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrqb2\" (UniqueName: \"kubernetes.io/projected/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-kube-api-access-nrqb2\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.316327 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-fernet-keys\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.316366 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-combined-ca-bundle\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.319722 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-config-data\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.319906 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrqb2\" (UniqueName: \"kubernetes.io/projected/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-kube-api-access-nrqb2\") pod \"keystone-cron-29412061-jvbbz\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:00 crc kubenswrapper[4912]: I1203 01:01:00.478253 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:01 crc kubenswrapper[4912]: W1203 01:01:01.009981 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65a4a770_16d8_4ea3_ac4a_9810ed0c2477.slice/crio-4ff67d4574e3056c364abd642b792abf8594fed79a69d13895e38d909fee3bcf WatchSource:0}: Error finding container 4ff67d4574e3056c364abd642b792abf8594fed79a69d13895e38d909fee3bcf: Status 404 returned error can't find the container with id 4ff67d4574e3056c364abd642b792abf8594fed79a69d13895e38d909fee3bcf Dec 03 01:01:01 crc kubenswrapper[4912]: I1203 01:01:01.017305 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412061-jvbbz"] Dec 03 01:01:01 crc kubenswrapper[4912]: I1203 01:01:01.486151 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412061-jvbbz" event={"ID":"65a4a770-16d8-4ea3-ac4a-9810ed0c2477","Type":"ContainerStarted","Data":"72e059dfe23e2f71c9c98b1680c63ec4b1967058ab24b67ed85cc7cc0d83bbb5"} Dec 03 01:01:01 crc kubenswrapper[4912]: I1203 01:01:01.488065 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412061-jvbbz" event={"ID":"65a4a770-16d8-4ea3-ac4a-9810ed0c2477","Type":"ContainerStarted","Data":"4ff67d4574e3056c364abd642b792abf8594fed79a69d13895e38d909fee3bcf"} Dec 03 01:01:01 crc kubenswrapper[4912]: I1203 01:01:01.517060 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29412061-jvbbz" podStartSLOduration=1.51704062 podStartE2EDuration="1.51704062s" podCreationTimestamp="2025-12-03 01:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:01:01.511915203 +0000 UTC m=+2247.153935773" watchObservedRunningTime="2025-12-03 01:01:01.51704062 +0000 UTC m=+2247.159061180" Dec 03 01:01:03 crc kubenswrapper[4912]: I1203 01:01:03.077887 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-k77b2"] Dec 03 01:01:03 crc kubenswrapper[4912]: I1203 01:01:03.092315 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-k77b2"] Dec 03 01:01:03 crc kubenswrapper[4912]: I1203 01:01:03.510935 4912 generic.go:334] "Generic (PLEG): container finished" podID="65a4a770-16d8-4ea3-ac4a-9810ed0c2477" containerID="72e059dfe23e2f71c9c98b1680c63ec4b1967058ab24b67ed85cc7cc0d83bbb5" exitCode=0 Dec 03 01:01:03 crc kubenswrapper[4912]: I1203 01:01:03.511030 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412061-jvbbz" event={"ID":"65a4a770-16d8-4ea3-ac4a-9810ed0c2477","Type":"ContainerDied","Data":"72e059dfe23e2f71c9c98b1680c63ec4b1967058ab24b67ed85cc7cc0d83bbb5"} Dec 03 01:01:03 crc kubenswrapper[4912]: E1203 01:01:03.551770 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65a4a770_16d8_4ea3_ac4a_9810ed0c2477.slice/crio-72e059dfe23e2f71c9c98b1680c63ec4b1967058ab24b67ed85cc7cc0d83bbb5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65a4a770_16d8_4ea3_ac4a_9810ed0c2477.slice/crio-conmon-72e059dfe23e2f71c9c98b1680c63ec4b1967058ab24b67ed85cc7cc0d83bbb5.scope\": RecentStats: unable to find data in memory cache]" Dec 03 01:01:04 crc kubenswrapper[4912]: I1203 01:01:04.596055 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="435cfdb9-676e-448f-bab9-d39e3cd95afd" path="/var/lib/kubelet/pods/435cfdb9-676e-448f-bab9-d39e3cd95afd/volumes" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.001702 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.033385 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-820b-account-create-update-khnl7"] Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.075798 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-820b-account-create-update-khnl7"] Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.144806 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-config-data\") pod \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.145339 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-combined-ca-bundle\") pod \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.145732 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-fernet-keys\") pod \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.145970 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrqb2\" (UniqueName: \"kubernetes.io/projected/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-kube-api-access-nrqb2\") pod \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\" (UID: \"65a4a770-16d8-4ea3-ac4a-9810ed0c2477\") " Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.152928 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "65a4a770-16d8-4ea3-ac4a-9810ed0c2477" (UID: "65a4a770-16d8-4ea3-ac4a-9810ed0c2477"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.153763 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-kube-api-access-nrqb2" (OuterVolumeSpecName: "kube-api-access-nrqb2") pod "65a4a770-16d8-4ea3-ac4a-9810ed0c2477" (UID: "65a4a770-16d8-4ea3-ac4a-9810ed0c2477"). InnerVolumeSpecName "kube-api-access-nrqb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.194778 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65a4a770-16d8-4ea3-ac4a-9810ed0c2477" (UID: "65a4a770-16d8-4ea3-ac4a-9810ed0c2477"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.244584 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-config-data" (OuterVolumeSpecName: "config-data") pod "65a4a770-16d8-4ea3-ac4a-9810ed0c2477" (UID: "65a4a770-16d8-4ea3-ac4a-9810ed0c2477"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.256838 4912 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.256895 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrqb2\" (UniqueName: \"kubernetes.io/projected/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-kube-api-access-nrqb2\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.256912 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.256921 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65a4a770-16d8-4ea3-ac4a-9810ed0c2477-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.540404 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412061-jvbbz" event={"ID":"65a4a770-16d8-4ea3-ac4a-9810ed0c2477","Type":"ContainerDied","Data":"4ff67d4574e3056c364abd642b792abf8594fed79a69d13895e38d909fee3bcf"} Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.540483 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ff67d4574e3056c364abd642b792abf8594fed79a69d13895e38d909fee3bcf" Dec 03 01:01:05 crc kubenswrapper[4912]: I1203 01:01:05.540531 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412061-jvbbz" Dec 03 01:01:06 crc kubenswrapper[4912]: I1203 01:01:06.588170 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d946f64a-de87-496f-992f-8922b2e061e2" path="/var/lib/kubelet/pods/d946f64a-de87-496f-992f-8922b2e061e2/volumes" Dec 03 01:01:10 crc kubenswrapper[4912]: I1203 01:01:10.039229 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-knjhl"] Dec 03 01:01:10 crc kubenswrapper[4912]: I1203 01:01:10.054295 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-knjhl"] Dec 03 01:01:10 crc kubenswrapper[4912]: I1203 01:01:10.590490 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc1d4411-e4c3-4107-90b4-77bd821e61dd" path="/var/lib/kubelet/pods/cc1d4411-e4c3-4107-90b4-77bd821e61dd/volumes" Dec 03 01:01:11 crc kubenswrapper[4912]: I1203 01:01:11.045211 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxxm4"] Dec 03 01:01:11 crc kubenswrapper[4912]: I1203 01:01:11.060820 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxxm4"] Dec 03 01:01:12 crc kubenswrapper[4912]: I1203 01:01:12.604899 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7bdcd2a-1372-4f41-a610-74547e28c9de" path="/var/lib/kubelet/pods/d7bdcd2a-1372-4f41-a610-74547e28c9de/volumes" Dec 03 01:01:18 crc kubenswrapper[4912]: I1203 01:01:18.078564 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:01:18 crc kubenswrapper[4912]: I1203 01:01:18.079577 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:01:18 crc kubenswrapper[4912]: I1203 01:01:18.079678 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:01:18 crc kubenswrapper[4912]: I1203 01:01:18.081249 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:01:18 crc kubenswrapper[4912]: I1203 01:01:18.081366 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" gracePeriod=600 Dec 03 01:01:18 crc kubenswrapper[4912]: E1203 01:01:18.231156 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:01:18 crc kubenswrapper[4912]: I1203 01:01:18.764863 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" exitCode=0 Dec 03 01:01:18 crc kubenswrapper[4912]: I1203 01:01:18.764941 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc"} Dec 03 01:01:18 crc kubenswrapper[4912]: I1203 01:01:18.765559 4912 scope.go:117] "RemoveContainer" containerID="87c39cfe128fba833e7df8ba7ab177c6dc006107e5cd523a76fe1db299c9b196" Dec 03 01:01:18 crc kubenswrapper[4912]: I1203 01:01:18.766891 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:01:18 crc kubenswrapper[4912]: E1203 01:01:18.767560 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:01:22 crc kubenswrapper[4912]: I1203 01:01:22.002731 4912 scope.go:117] "RemoveContainer" containerID="0c0f4f27ed268c3a0f014812eeaf7bbbbb75ef2670c50aefd3e025b54b275d4d" Dec 03 01:01:22 crc kubenswrapper[4912]: I1203 01:01:22.084589 4912 scope.go:117] "RemoveContainer" containerID="e7fc4eb2633528ebde3b03fb7b350e58a1a0a23594a6494c01b2c7016451eba8" Dec 03 01:01:22 crc kubenswrapper[4912]: I1203 01:01:22.113116 4912 scope.go:117] "RemoveContainer" containerID="6d55544f43c33af5471ca5230284ff68432e49975670653a34aebe15b04d759b" Dec 03 01:01:22 crc kubenswrapper[4912]: I1203 01:01:22.205865 4912 scope.go:117] "RemoveContainer" containerID="b94ce68ff5e0303a26a0f97378464cf51d8756cd2a1d9681f8f4f5cfca7449f7" Dec 03 01:01:22 crc kubenswrapper[4912]: I1203 01:01:22.231375 4912 scope.go:117] "RemoveContainer" containerID="d8d0315b779492e8da1aad0ac6bd679a8db367aa217f9b35baff966ae1e97134" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.046749 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzjt"] Dec 03 01:01:23 crc kubenswrapper[4912]: E1203 01:01:23.047324 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65a4a770-16d8-4ea3-ac4a-9810ed0c2477" containerName="keystone-cron" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.047337 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="65a4a770-16d8-4ea3-ac4a-9810ed0c2477" containerName="keystone-cron" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.047571 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="65a4a770-16d8-4ea3-ac4a-9810ed0c2477" containerName="keystone-cron" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.049361 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.072086 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzjt"] Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.107101 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-utilities\") pod \"redhat-marketplace-4vzjt\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.107462 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-catalog-content\") pod \"redhat-marketplace-4vzjt\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.107774 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr2xx\" (UniqueName: \"kubernetes.io/projected/b020c274-c64b-4c11-8855-82918bbaf6da-kube-api-access-tr2xx\") pod \"redhat-marketplace-4vzjt\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.210312 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr2xx\" (UniqueName: \"kubernetes.io/projected/b020c274-c64b-4c11-8855-82918bbaf6da-kube-api-access-tr2xx\") pod \"redhat-marketplace-4vzjt\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.210503 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-utilities\") pod \"redhat-marketplace-4vzjt\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.210550 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-catalog-content\") pod \"redhat-marketplace-4vzjt\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.211458 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-utilities\") pod \"redhat-marketplace-4vzjt\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.211563 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-catalog-content\") pod \"redhat-marketplace-4vzjt\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.231640 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr2xx\" (UniqueName: \"kubernetes.io/projected/b020c274-c64b-4c11-8855-82918bbaf6da-kube-api-access-tr2xx\") pod \"redhat-marketplace-4vzjt\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.377186 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:23 crc kubenswrapper[4912]: I1203 01:01:23.977162 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzjt"] Dec 03 01:01:24 crc kubenswrapper[4912]: I1203 01:01:24.865156 4912 generic.go:334] "Generic (PLEG): container finished" podID="b020c274-c64b-4c11-8855-82918bbaf6da" containerID="0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d" exitCode=0 Dec 03 01:01:24 crc kubenswrapper[4912]: I1203 01:01:24.865285 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzjt" event={"ID":"b020c274-c64b-4c11-8855-82918bbaf6da","Type":"ContainerDied","Data":"0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d"} Dec 03 01:01:24 crc kubenswrapper[4912]: I1203 01:01:24.865618 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzjt" event={"ID":"b020c274-c64b-4c11-8855-82918bbaf6da","Type":"ContainerStarted","Data":"031edf84f274cabedf514c238e1fe0488798970c329bf646804b139a6835688f"} Dec 03 01:01:29 crc kubenswrapper[4912]: I1203 01:01:29.572650 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:01:29 crc kubenswrapper[4912]: E1203 01:01:29.574198 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:01:29 crc kubenswrapper[4912]: I1203 01:01:29.942763 4912 generic.go:334] "Generic (PLEG): container finished" podID="b020c274-c64b-4c11-8855-82918bbaf6da" containerID="4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5" exitCode=0 Dec 03 01:01:29 crc kubenswrapper[4912]: I1203 01:01:29.943245 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzjt" event={"ID":"b020c274-c64b-4c11-8855-82918bbaf6da","Type":"ContainerDied","Data":"4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5"} Dec 03 01:01:31 crc kubenswrapper[4912]: I1203 01:01:31.984489 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzjt" event={"ID":"b020c274-c64b-4c11-8855-82918bbaf6da","Type":"ContainerStarted","Data":"f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66"} Dec 03 01:01:32 crc kubenswrapper[4912]: I1203 01:01:32.036627 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4vzjt" podStartSLOduration=2.986567715 podStartE2EDuration="9.036599937s" podCreationTimestamp="2025-12-03 01:01:23 +0000 UTC" firstStartedPulling="2025-12-03 01:01:24.866624578 +0000 UTC m=+2270.508645148" lastFinishedPulling="2025-12-03 01:01:30.9166568 +0000 UTC m=+2276.558677370" observedRunningTime="2025-12-03 01:01:32.010999842 +0000 UTC m=+2277.653020402" watchObservedRunningTime="2025-12-03 01:01:32.036599937 +0000 UTC m=+2277.678620497" Dec 03 01:01:33 crc kubenswrapper[4912]: I1203 01:01:33.377758 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:33 crc kubenswrapper[4912]: I1203 01:01:33.377848 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:34 crc kubenswrapper[4912]: I1203 01:01:34.429184 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-4vzjt" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" containerName="registry-server" probeResult="failure" output=< Dec 03 01:01:34 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 01:01:34 crc kubenswrapper[4912]: > Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.469931 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kljgx"] Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.476477 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.489730 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kljgx"] Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.621329 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-utilities\") pod \"certified-operators-kljgx\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.621647 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-catalog-content\") pod \"certified-operators-kljgx\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.621828 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz7mg\" (UniqueName: \"kubernetes.io/projected/2d648983-0ba1-4555-a918-62b4c2a23881-kube-api-access-jz7mg\") pod \"certified-operators-kljgx\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.724520 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz7mg\" (UniqueName: \"kubernetes.io/projected/2d648983-0ba1-4555-a918-62b4c2a23881-kube-api-access-jz7mg\") pod \"certified-operators-kljgx\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.724701 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-utilities\") pod \"certified-operators-kljgx\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.724875 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-catalog-content\") pod \"certified-operators-kljgx\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.726291 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-catalog-content\") pod \"certified-operators-kljgx\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.726483 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-utilities\") pod \"certified-operators-kljgx\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.751175 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz7mg\" (UniqueName: \"kubernetes.io/projected/2d648983-0ba1-4555-a918-62b4c2a23881-kube-api-access-jz7mg\") pod \"certified-operators-kljgx\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:38 crc kubenswrapper[4912]: I1203 01:01:38.841664 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:39 crc kubenswrapper[4912]: I1203 01:01:39.369161 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kljgx"] Dec 03 01:01:39 crc kubenswrapper[4912]: W1203 01:01:39.372693 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d648983_0ba1_4555_a918_62b4c2a23881.slice/crio-3b895b55347e6dd9ca07b57f7b95e28dcebf6bc10a6d3b135d13cd9b86ff6eda WatchSource:0}: Error finding container 3b895b55347e6dd9ca07b57f7b95e28dcebf6bc10a6d3b135d13cd9b86ff6eda: Status 404 returned error can't find the container with id 3b895b55347e6dd9ca07b57f7b95e28dcebf6bc10a6d3b135d13cd9b86ff6eda Dec 03 01:01:40 crc kubenswrapper[4912]: I1203 01:01:40.107214 4912 generic.go:334] "Generic (PLEG): container finished" podID="2d648983-0ba1-4555-a918-62b4c2a23881" containerID="e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947" exitCode=0 Dec 03 01:01:40 crc kubenswrapper[4912]: I1203 01:01:40.107415 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kljgx" event={"ID":"2d648983-0ba1-4555-a918-62b4c2a23881","Type":"ContainerDied","Data":"e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947"} Dec 03 01:01:40 crc kubenswrapper[4912]: I1203 01:01:40.107546 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kljgx" event={"ID":"2d648983-0ba1-4555-a918-62b4c2a23881","Type":"ContainerStarted","Data":"3b895b55347e6dd9ca07b57f7b95e28dcebf6bc10a6d3b135d13cd9b86ff6eda"} Dec 03 01:01:42 crc kubenswrapper[4912]: I1203 01:01:42.148714 4912 generic.go:334] "Generic (PLEG): container finished" podID="2d648983-0ba1-4555-a918-62b4c2a23881" containerID="c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d" exitCode=0 Dec 03 01:01:42 crc kubenswrapper[4912]: I1203 01:01:42.148898 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kljgx" event={"ID":"2d648983-0ba1-4555-a918-62b4c2a23881","Type":"ContainerDied","Data":"c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d"} Dec 03 01:01:42 crc kubenswrapper[4912]: I1203 01:01:42.576708 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:01:42 crc kubenswrapper[4912]: E1203 01:01:42.577970 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:01:43 crc kubenswrapper[4912]: I1203 01:01:43.163107 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kljgx" event={"ID":"2d648983-0ba1-4555-a918-62b4c2a23881","Type":"ContainerStarted","Data":"9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c"} Dec 03 01:01:43 crc kubenswrapper[4912]: I1203 01:01:43.185757 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kljgx" podStartSLOduration=2.562532888 podStartE2EDuration="5.185734035s" podCreationTimestamp="2025-12-03 01:01:38 +0000 UTC" firstStartedPulling="2025-12-03 01:01:40.109699674 +0000 UTC m=+2285.751720264" lastFinishedPulling="2025-12-03 01:01:42.732900861 +0000 UTC m=+2288.374921411" observedRunningTime="2025-12-03 01:01:43.182342143 +0000 UTC m=+2288.824362733" watchObservedRunningTime="2025-12-03 01:01:43.185734035 +0000 UTC m=+2288.827754605" Dec 03 01:01:43 crc kubenswrapper[4912]: I1203 01:01:43.452306 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:43 crc kubenswrapper[4912]: I1203 01:01:43.515005 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:44 crc kubenswrapper[4912]: I1203 01:01:44.944570 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzjt"] Dec 03 01:01:45 crc kubenswrapper[4912]: I1203 01:01:45.189826 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4vzjt" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" containerName="registry-server" containerID="cri-o://f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66" gracePeriod=2 Dec 03 01:01:45 crc kubenswrapper[4912]: I1203 01:01:45.851363 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:45 crc kubenswrapper[4912]: I1203 01:01:45.942316 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr2xx\" (UniqueName: \"kubernetes.io/projected/b020c274-c64b-4c11-8855-82918bbaf6da-kube-api-access-tr2xx\") pod \"b020c274-c64b-4c11-8855-82918bbaf6da\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " Dec 03 01:01:45 crc kubenswrapper[4912]: I1203 01:01:45.942515 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-utilities\") pod \"b020c274-c64b-4c11-8855-82918bbaf6da\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " Dec 03 01:01:45 crc kubenswrapper[4912]: I1203 01:01:45.942904 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-catalog-content\") pod \"b020c274-c64b-4c11-8855-82918bbaf6da\" (UID: \"b020c274-c64b-4c11-8855-82918bbaf6da\") " Dec 03 01:01:45 crc kubenswrapper[4912]: I1203 01:01:45.944197 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-utilities" (OuterVolumeSpecName: "utilities") pod "b020c274-c64b-4c11-8855-82918bbaf6da" (UID: "b020c274-c64b-4c11-8855-82918bbaf6da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:01:45 crc kubenswrapper[4912]: I1203 01:01:45.952833 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b020c274-c64b-4c11-8855-82918bbaf6da-kube-api-access-tr2xx" (OuterVolumeSpecName: "kube-api-access-tr2xx") pod "b020c274-c64b-4c11-8855-82918bbaf6da" (UID: "b020c274-c64b-4c11-8855-82918bbaf6da"). InnerVolumeSpecName "kube-api-access-tr2xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:01:45 crc kubenswrapper[4912]: I1203 01:01:45.958949 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b020c274-c64b-4c11-8855-82918bbaf6da" (UID: "b020c274-c64b-4c11-8855-82918bbaf6da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.046262 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.046318 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr2xx\" (UniqueName: \"kubernetes.io/projected/b020c274-c64b-4c11-8855-82918bbaf6da-kube-api-access-tr2xx\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.046343 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b020c274-c64b-4c11-8855-82918bbaf6da-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.206481 4912 generic.go:334] "Generic (PLEG): container finished" podID="b020c274-c64b-4c11-8855-82918bbaf6da" containerID="f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66" exitCode=0 Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.206538 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzjt" event={"ID":"b020c274-c64b-4c11-8855-82918bbaf6da","Type":"ContainerDied","Data":"f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66"} Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.206588 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzjt" event={"ID":"b020c274-c64b-4c11-8855-82918bbaf6da","Type":"ContainerDied","Data":"031edf84f274cabedf514c238e1fe0488798970c329bf646804b139a6835688f"} Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.206606 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vzjt" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.206610 4912 scope.go:117] "RemoveContainer" containerID="f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.265076 4912 scope.go:117] "RemoveContainer" containerID="4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.277934 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzjt"] Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.290772 4912 scope.go:117] "RemoveContainer" containerID="0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.291607 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzjt"] Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.345951 4912 scope.go:117] "RemoveContainer" containerID="f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66" Dec 03 01:01:46 crc kubenswrapper[4912]: E1203 01:01:46.346614 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66\": container with ID starting with f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66 not found: ID does not exist" containerID="f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.346652 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66"} err="failed to get container status \"f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66\": rpc error: code = NotFound desc = could not find container \"f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66\": container with ID starting with f73e8bd32e2c2bd83e0444796fc112104ab4e8d8f4e6e09ee63c7f2c0a07ec66 not found: ID does not exist" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.346680 4912 scope.go:117] "RemoveContainer" containerID="4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5" Dec 03 01:01:46 crc kubenswrapper[4912]: E1203 01:01:46.347141 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5\": container with ID starting with 4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5 not found: ID does not exist" containerID="4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.347167 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5"} err="failed to get container status \"4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5\": rpc error: code = NotFound desc = could not find container \"4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5\": container with ID starting with 4b308ac168e4e87175454a6193aa8050c9ab23bec1124dbac59040daee52dbf5 not found: ID does not exist" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.347185 4912 scope.go:117] "RemoveContainer" containerID="0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d" Dec 03 01:01:46 crc kubenswrapper[4912]: E1203 01:01:46.347555 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d\": container with ID starting with 0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d not found: ID does not exist" containerID="0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.347578 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d"} err="failed to get container status \"0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d\": rpc error: code = NotFound desc = could not find container \"0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d\": container with ID starting with 0770ebe8c75086706179be5c119221af6848feb105e0e32e5e721b86975fb11d not found: ID does not exist" Dec 03 01:01:46 crc kubenswrapper[4912]: I1203 01:01:46.583670 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" path="/var/lib/kubelet/pods/b020c274-c64b-4c11-8855-82918bbaf6da/volumes" Dec 03 01:01:48 crc kubenswrapper[4912]: I1203 01:01:48.842100 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:48 crc kubenswrapper[4912]: I1203 01:01:48.842562 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:48 crc kubenswrapper[4912]: I1203 01:01:48.906505 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:49 crc kubenswrapper[4912]: I1203 01:01:49.338583 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:50 crc kubenswrapper[4912]: I1203 01:01:50.254062 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kljgx"] Dec 03 01:01:51 crc kubenswrapper[4912]: I1203 01:01:51.304408 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kljgx" podUID="2d648983-0ba1-4555-a918-62b4c2a23881" containerName="registry-server" containerID="cri-o://9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c" gracePeriod=2 Dec 03 01:01:51 crc kubenswrapper[4912]: I1203 01:01:51.916549 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.030123 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz7mg\" (UniqueName: \"kubernetes.io/projected/2d648983-0ba1-4555-a918-62b4c2a23881-kube-api-access-jz7mg\") pod \"2d648983-0ba1-4555-a918-62b4c2a23881\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.030211 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-catalog-content\") pod \"2d648983-0ba1-4555-a918-62b4c2a23881\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.030250 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-utilities\") pod \"2d648983-0ba1-4555-a918-62b4c2a23881\" (UID: \"2d648983-0ba1-4555-a918-62b4c2a23881\") " Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.031279 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-utilities" (OuterVolumeSpecName: "utilities") pod "2d648983-0ba1-4555-a918-62b4c2a23881" (UID: "2d648983-0ba1-4555-a918-62b4c2a23881"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.040637 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d648983-0ba1-4555-a918-62b4c2a23881-kube-api-access-jz7mg" (OuterVolumeSpecName: "kube-api-access-jz7mg") pod "2d648983-0ba1-4555-a918-62b4c2a23881" (UID: "2d648983-0ba1-4555-a918-62b4c2a23881"). InnerVolumeSpecName "kube-api-access-jz7mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.094586 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d648983-0ba1-4555-a918-62b4c2a23881" (UID: "2d648983-0ba1-4555-a918-62b4c2a23881"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.133754 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz7mg\" (UniqueName: \"kubernetes.io/projected/2d648983-0ba1-4555-a918-62b4c2a23881-kube-api-access-jz7mg\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.133807 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.133824 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d648983-0ba1-4555-a918-62b4c2a23881-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.314870 4912 generic.go:334] "Generic (PLEG): container finished" podID="2d648983-0ba1-4555-a918-62b4c2a23881" containerID="9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c" exitCode=0 Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.314930 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kljgx" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.314934 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kljgx" event={"ID":"2d648983-0ba1-4555-a918-62b4c2a23881","Type":"ContainerDied","Data":"9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c"} Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.315121 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kljgx" event={"ID":"2d648983-0ba1-4555-a918-62b4c2a23881","Type":"ContainerDied","Data":"3b895b55347e6dd9ca07b57f7b95e28dcebf6bc10a6d3b135d13cd9b86ff6eda"} Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.315143 4912 scope.go:117] "RemoveContainer" containerID="9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.335665 4912 scope.go:117] "RemoveContainer" containerID="c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.360148 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kljgx"] Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.378108 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kljgx"] Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.378928 4912 scope.go:117] "RemoveContainer" containerID="e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.437621 4912 scope.go:117] "RemoveContainer" containerID="9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c" Dec 03 01:01:52 crc kubenswrapper[4912]: E1203 01:01:52.443740 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c\": container with ID starting with 9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c not found: ID does not exist" containerID="9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.443796 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c"} err="failed to get container status \"9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c\": rpc error: code = NotFound desc = could not find container \"9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c\": container with ID starting with 9ea4d56a72e57fbf4f0882054f19b2b214e664c8b32d684866918a30d2ca7e3c not found: ID does not exist" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.443832 4912 scope.go:117] "RemoveContainer" containerID="c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d" Dec 03 01:01:52 crc kubenswrapper[4912]: E1203 01:01:52.445893 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d\": container with ID starting with c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d not found: ID does not exist" containerID="c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.445944 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d"} err="failed to get container status \"c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d\": rpc error: code = NotFound desc = could not find container \"c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d\": container with ID starting with c55fe7dd588628665d4213799e9f02699e253d9d9ae86c62d5984b9f3136cd3d not found: ID does not exist" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.445978 4912 scope.go:117] "RemoveContainer" containerID="e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947" Dec 03 01:01:52 crc kubenswrapper[4912]: E1203 01:01:52.448730 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947\": container with ID starting with e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947 not found: ID does not exist" containerID="e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.448762 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947"} err="failed to get container status \"e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947\": rpc error: code = NotFound desc = could not find container \"e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947\": container with ID starting with e997df4d491025d1174e3f002fed5a3610f361de238adab3eb8be0167e87a947 not found: ID does not exist" Dec 03 01:01:52 crc kubenswrapper[4912]: I1203 01:01:52.595828 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d648983-0ba1-4555-a918-62b4c2a23881" path="/var/lib/kubelet/pods/2d648983-0ba1-4555-a918-62b4c2a23881/volumes" Dec 03 01:01:56 crc kubenswrapper[4912]: I1203 01:01:56.572910 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:01:56 crc kubenswrapper[4912]: E1203 01:01:56.574222 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:01:58 crc kubenswrapper[4912]: I1203 01:01:58.072539 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-sgfjt"] Dec 03 01:01:58 crc kubenswrapper[4912]: I1203 01:01:58.090419 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-sgfjt"] Dec 03 01:01:58 crc kubenswrapper[4912]: I1203 01:01:58.592591 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9" path="/var/lib/kubelet/pods/a9684c3e-2b6a-4f7a-ae00-e323c6dc4fe9/volumes" Dec 03 01:02:04 crc kubenswrapper[4912]: I1203 01:02:04.492465 4912 generic.go:334] "Generic (PLEG): container finished" podID="0908ef60-cdec-46de-beb6-167295e04796" containerID="5c9d9eff1acf7d358d730edc5e20382faa3aca1799e3ace928e84fadea14bc80" exitCode=0 Dec 03 01:02:04 crc kubenswrapper[4912]: I1203 01:02:04.492604 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" event={"ID":"0908ef60-cdec-46de-beb6-167295e04796","Type":"ContainerDied","Data":"5c9d9eff1acf7d358d730edc5e20382faa3aca1799e3ace928e84fadea14bc80"} Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.068870 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.117680 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-inventory\") pod \"0908ef60-cdec-46de-beb6-167295e04796\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.118028 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfq8n\" (UniqueName: \"kubernetes.io/projected/0908ef60-cdec-46de-beb6-167295e04796-kube-api-access-jfq8n\") pod \"0908ef60-cdec-46de-beb6-167295e04796\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.118079 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-ssh-key\") pod \"0908ef60-cdec-46de-beb6-167295e04796\" (UID: \"0908ef60-cdec-46de-beb6-167295e04796\") " Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.137004 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0908ef60-cdec-46de-beb6-167295e04796-kube-api-access-jfq8n" (OuterVolumeSpecName: "kube-api-access-jfq8n") pod "0908ef60-cdec-46de-beb6-167295e04796" (UID: "0908ef60-cdec-46de-beb6-167295e04796"). InnerVolumeSpecName "kube-api-access-jfq8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.161823 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-inventory" (OuterVolumeSpecName: "inventory") pod "0908ef60-cdec-46de-beb6-167295e04796" (UID: "0908ef60-cdec-46de-beb6-167295e04796"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.163203 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0908ef60-cdec-46de-beb6-167295e04796" (UID: "0908ef60-cdec-46de-beb6-167295e04796"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.221124 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.221167 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0908ef60-cdec-46de-beb6-167295e04796-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.221180 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfq8n\" (UniqueName: \"kubernetes.io/projected/0908ef60-cdec-46de-beb6-167295e04796-kube-api-access-jfq8n\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.529077 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" event={"ID":"0908ef60-cdec-46de-beb6-167295e04796","Type":"ContainerDied","Data":"90ba08b2e7f09f42cd7b2937cb60c5578c912972f6d65914424bf176d7d31bc2"} Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.529120 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90ba08b2e7f09f42cd7b2937cb60c5578c912972f6d65914424bf176d7d31bc2" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.529210 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.631564 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zmwqm"] Dec 03 01:02:06 crc kubenswrapper[4912]: E1203 01:02:06.631973 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0908ef60-cdec-46de-beb6-167295e04796" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.631992 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0908ef60-cdec-46de-beb6-167295e04796" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:02:06 crc kubenswrapper[4912]: E1203 01:02:06.632012 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" containerName="registry-server" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.632019 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" containerName="registry-server" Dec 03 01:02:06 crc kubenswrapper[4912]: E1203 01:02:06.632031 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d648983-0ba1-4555-a918-62b4c2a23881" containerName="extract-content" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.632038 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d648983-0ba1-4555-a918-62b4c2a23881" containerName="extract-content" Dec 03 01:02:06 crc kubenswrapper[4912]: E1203 01:02:06.632046 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" containerName="extract-utilities" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.632054 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" containerName="extract-utilities" Dec 03 01:02:06 crc kubenswrapper[4912]: E1203 01:02:06.632091 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" containerName="extract-content" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.632097 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" containerName="extract-content" Dec 03 01:02:06 crc kubenswrapper[4912]: E1203 01:02:06.632111 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d648983-0ba1-4555-a918-62b4c2a23881" containerName="extract-utilities" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.632117 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d648983-0ba1-4555-a918-62b4c2a23881" containerName="extract-utilities" Dec 03 01:02:06 crc kubenswrapper[4912]: E1203 01:02:06.632137 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d648983-0ba1-4555-a918-62b4c2a23881" containerName="registry-server" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.632143 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d648983-0ba1-4555-a918-62b4c2a23881" containerName="registry-server" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.632339 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0908ef60-cdec-46de-beb6-167295e04796" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.632353 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="b020c274-c64b-4c11-8855-82918bbaf6da" containerName="registry-server" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.632371 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d648983-0ba1-4555-a918-62b4c2a23881" containerName="registry-server" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.634874 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.641373 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.641416 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.641667 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.643911 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.649092 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zmwqm"] Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.731709 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zmwqm\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.731881 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsfpx\" (UniqueName: \"kubernetes.io/projected/066dd271-4382-440a-9968-c02d9393cbd1-kube-api-access-rsfpx\") pod \"ssh-known-hosts-edpm-deployment-zmwqm\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.731931 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zmwqm\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.835179 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsfpx\" (UniqueName: \"kubernetes.io/projected/066dd271-4382-440a-9968-c02d9393cbd1-kube-api-access-rsfpx\") pod \"ssh-known-hosts-edpm-deployment-zmwqm\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.835271 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zmwqm\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.835390 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zmwqm\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.841127 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-zmwqm\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.843285 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-zmwqm\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.859606 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsfpx\" (UniqueName: \"kubernetes.io/projected/066dd271-4382-440a-9968-c02d9393cbd1-kube-api-access-rsfpx\") pod \"ssh-known-hosts-edpm-deployment-zmwqm\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:06 crc kubenswrapper[4912]: I1203 01:02:06.975271 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:07 crc kubenswrapper[4912]: I1203 01:02:07.605441 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zmwqm"] Dec 03 01:02:08 crc kubenswrapper[4912]: I1203 01:02:08.557406 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" event={"ID":"066dd271-4382-440a-9968-c02d9393cbd1","Type":"ContainerStarted","Data":"ab55aab0ad54779688c3948006ebfb6e69d5a5689c75f8b47a6eb20424b7df3a"} Dec 03 01:02:08 crc kubenswrapper[4912]: I1203 01:02:08.557995 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" event={"ID":"066dd271-4382-440a-9968-c02d9393cbd1","Type":"ContainerStarted","Data":"06bc23e85bbbfcc5bb3ab60b81415b3f8b08bf8cac1d99b6af61b581f84d59fd"} Dec 03 01:02:08 crc kubenswrapper[4912]: I1203 01:02:08.571934 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:02:08 crc kubenswrapper[4912]: E1203 01:02:08.572285 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:02:08 crc kubenswrapper[4912]: I1203 01:02:08.592748 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" podStartSLOduration=1.968364825 podStartE2EDuration="2.592718785s" podCreationTimestamp="2025-12-03 01:02:06 +0000 UTC" firstStartedPulling="2025-12-03 01:02:07.630842956 +0000 UTC m=+2313.272863516" lastFinishedPulling="2025-12-03 01:02:08.255196916 +0000 UTC m=+2313.897217476" observedRunningTime="2025-12-03 01:02:08.578567696 +0000 UTC m=+2314.220588286" watchObservedRunningTime="2025-12-03 01:02:08.592718785 +0000 UTC m=+2314.234739345" Dec 03 01:02:17 crc kubenswrapper[4912]: I1203 01:02:17.709243 4912 generic.go:334] "Generic (PLEG): container finished" podID="066dd271-4382-440a-9968-c02d9393cbd1" containerID="ab55aab0ad54779688c3948006ebfb6e69d5a5689c75f8b47a6eb20424b7df3a" exitCode=0 Dec 03 01:02:17 crc kubenswrapper[4912]: I1203 01:02:17.709311 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" event={"ID":"066dd271-4382-440a-9968-c02d9393cbd1","Type":"ContainerDied","Data":"ab55aab0ad54779688c3948006ebfb6e69d5a5689c75f8b47a6eb20424b7df3a"} Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.296004 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.397106 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-inventory-0\") pod \"066dd271-4382-440a-9968-c02d9393cbd1\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.397205 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-ssh-key-openstack-edpm-ipam\") pod \"066dd271-4382-440a-9968-c02d9393cbd1\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.398037 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsfpx\" (UniqueName: \"kubernetes.io/projected/066dd271-4382-440a-9968-c02d9393cbd1-kube-api-access-rsfpx\") pod \"066dd271-4382-440a-9968-c02d9393cbd1\" (UID: \"066dd271-4382-440a-9968-c02d9393cbd1\") " Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.409225 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/066dd271-4382-440a-9968-c02d9393cbd1-kube-api-access-rsfpx" (OuterVolumeSpecName: "kube-api-access-rsfpx") pod "066dd271-4382-440a-9968-c02d9393cbd1" (UID: "066dd271-4382-440a-9968-c02d9393cbd1"). InnerVolumeSpecName "kube-api-access-rsfpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.429086 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "066dd271-4382-440a-9968-c02d9393cbd1" (UID: "066dd271-4382-440a-9968-c02d9393cbd1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.429539 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "066dd271-4382-440a-9968-c02d9393cbd1" (UID: "066dd271-4382-440a-9968-c02d9393cbd1"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.501281 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsfpx\" (UniqueName: \"kubernetes.io/projected/066dd271-4382-440a-9968-c02d9393cbd1-kube-api-access-rsfpx\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.501324 4912 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.501337 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/066dd271-4382-440a-9968-c02d9393cbd1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.572077 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:02:19 crc kubenswrapper[4912]: E1203 01:02:19.572500 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.742688 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" event={"ID":"066dd271-4382-440a-9968-c02d9393cbd1","Type":"ContainerDied","Data":"06bc23e85bbbfcc5bb3ab60b81415b3f8b08bf8cac1d99b6af61b581f84d59fd"} Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.742732 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06bc23e85bbbfcc5bb3ab60b81415b3f8b08bf8cac1d99b6af61b581f84d59fd" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.742761 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-zmwqm" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.860115 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96"] Dec 03 01:02:19 crc kubenswrapper[4912]: E1203 01:02:19.860599 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="066dd271-4382-440a-9968-c02d9393cbd1" containerName="ssh-known-hosts-edpm-deployment" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.860617 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="066dd271-4382-440a-9968-c02d9393cbd1" containerName="ssh-known-hosts-edpm-deployment" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.860857 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="066dd271-4382-440a-9968-c02d9393cbd1" containerName="ssh-known-hosts-edpm-deployment" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.869547 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.880931 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96"] Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.883776 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.885212 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.885421 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.886021 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.913272 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-htl96\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.913327 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kjjs\" (UniqueName: \"kubernetes.io/projected/ef3338b7-de78-4aea-8f3b-809db2358f0e-kube-api-access-7kjjs\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-htl96\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:19 crc kubenswrapper[4912]: I1203 01:02:19.913352 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-htl96\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:20 crc kubenswrapper[4912]: I1203 01:02:20.014496 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kjjs\" (UniqueName: \"kubernetes.io/projected/ef3338b7-de78-4aea-8f3b-809db2358f0e-kube-api-access-7kjjs\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-htl96\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:20 crc kubenswrapper[4912]: I1203 01:02:20.014550 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-htl96\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:20 crc kubenswrapper[4912]: I1203 01:02:20.014701 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-htl96\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:20 crc kubenswrapper[4912]: I1203 01:02:20.019517 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-htl96\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:20 crc kubenswrapper[4912]: I1203 01:02:20.019907 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-htl96\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:20 crc kubenswrapper[4912]: I1203 01:02:20.034903 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kjjs\" (UniqueName: \"kubernetes.io/projected/ef3338b7-de78-4aea-8f3b-809db2358f0e-kube-api-access-7kjjs\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-htl96\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:20 crc kubenswrapper[4912]: I1203 01:02:20.189319 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:20 crc kubenswrapper[4912]: I1203 01:02:20.817174 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96"] Dec 03 01:02:21 crc kubenswrapper[4912]: I1203 01:02:21.774867 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" event={"ID":"ef3338b7-de78-4aea-8f3b-809db2358f0e","Type":"ContainerStarted","Data":"cc3096107788de126604dc147eed2f7c40312aa523a426f83a7f8352aab9926b"} Dec 03 01:02:21 crc kubenswrapper[4912]: I1203 01:02:21.775308 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" event={"ID":"ef3338b7-de78-4aea-8f3b-809db2358f0e","Type":"ContainerStarted","Data":"c70658e566997d960e68cfe7dc31410f24218bc8a083d375fa4bc44a0681af39"} Dec 03 01:02:21 crc kubenswrapper[4912]: I1203 01:02:21.801832 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" podStartSLOduration=2.221755878 podStartE2EDuration="2.801809854s" podCreationTimestamp="2025-12-03 01:02:19 +0000 UTC" firstStartedPulling="2025-12-03 01:02:20.820835354 +0000 UTC m=+2326.462855914" lastFinishedPulling="2025-12-03 01:02:21.40088933 +0000 UTC m=+2327.042909890" observedRunningTime="2025-12-03 01:02:21.794629691 +0000 UTC m=+2327.436650291" watchObservedRunningTime="2025-12-03 01:02:21.801809854 +0000 UTC m=+2327.443830424" Dec 03 01:02:22 crc kubenswrapper[4912]: I1203 01:02:22.403935 4912 scope.go:117] "RemoveContainer" containerID="c92792f6d2e7d39a1ce700c1e84f032ce8f7da2035aeca94f08b7a6de4a897ca" Dec 03 01:02:31 crc kubenswrapper[4912]: I1203 01:02:31.917414 4912 generic.go:334] "Generic (PLEG): container finished" podID="ef3338b7-de78-4aea-8f3b-809db2358f0e" containerID="cc3096107788de126604dc147eed2f7c40312aa523a426f83a7f8352aab9926b" exitCode=0 Dec 03 01:02:31 crc kubenswrapper[4912]: I1203 01:02:31.918092 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" event={"ID":"ef3338b7-de78-4aea-8f3b-809db2358f0e","Type":"ContainerDied","Data":"cc3096107788de126604dc147eed2f7c40312aa523a426f83a7f8352aab9926b"} Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.570605 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.575123 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:02:33 crc kubenswrapper[4912]: E1203 01:02:33.575681 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.697548 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kjjs\" (UniqueName: \"kubernetes.io/projected/ef3338b7-de78-4aea-8f3b-809db2358f0e-kube-api-access-7kjjs\") pod \"ef3338b7-de78-4aea-8f3b-809db2358f0e\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.697954 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-ssh-key\") pod \"ef3338b7-de78-4aea-8f3b-809db2358f0e\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.698223 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-inventory\") pod \"ef3338b7-de78-4aea-8f3b-809db2358f0e\" (UID: \"ef3338b7-de78-4aea-8f3b-809db2358f0e\") " Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.722257 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef3338b7-de78-4aea-8f3b-809db2358f0e-kube-api-access-7kjjs" (OuterVolumeSpecName: "kube-api-access-7kjjs") pod "ef3338b7-de78-4aea-8f3b-809db2358f0e" (UID: "ef3338b7-de78-4aea-8f3b-809db2358f0e"). InnerVolumeSpecName "kube-api-access-7kjjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.729980 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ef3338b7-de78-4aea-8f3b-809db2358f0e" (UID: "ef3338b7-de78-4aea-8f3b-809db2358f0e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.738299 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-inventory" (OuterVolumeSpecName: "inventory") pod "ef3338b7-de78-4aea-8f3b-809db2358f0e" (UID: "ef3338b7-de78-4aea-8f3b-809db2358f0e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.801776 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kjjs\" (UniqueName: \"kubernetes.io/projected/ef3338b7-de78-4aea-8f3b-809db2358f0e-kube-api-access-7kjjs\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.801861 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.801881 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef3338b7-de78-4aea-8f3b-809db2358f0e-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.950786 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" event={"ID":"ef3338b7-de78-4aea-8f3b-809db2358f0e","Type":"ContainerDied","Data":"c70658e566997d960e68cfe7dc31410f24218bc8a083d375fa4bc44a0681af39"} Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.950853 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c70658e566997d960e68cfe7dc31410f24218bc8a083d375fa4bc44a0681af39" Dec 03 01:02:33 crc kubenswrapper[4912]: I1203 01:02:33.951518 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.065873 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx"] Dec 03 01:02:34 crc kubenswrapper[4912]: E1203 01:02:34.066574 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef3338b7-de78-4aea-8f3b-809db2358f0e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.066596 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef3338b7-de78-4aea-8f3b-809db2358f0e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.066867 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef3338b7-de78-4aea-8f3b-809db2358f0e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.067911 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.071082 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.071197 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.071569 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.072050 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.096979 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx"] Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.210396 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-25csx\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.210536 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-25csx\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.210724 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xzs9\" (UniqueName: \"kubernetes.io/projected/afa950b8-ea20-43c7-bb99-0aa5476a6d83-kube-api-access-5xzs9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-25csx\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.313681 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-25csx\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.313775 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-25csx\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.313850 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xzs9\" (UniqueName: \"kubernetes.io/projected/afa950b8-ea20-43c7-bb99-0aa5476a6d83-kube-api-access-5xzs9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-25csx\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.320033 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-25csx\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.324281 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-25csx\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.349047 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xzs9\" (UniqueName: \"kubernetes.io/projected/afa950b8-ea20-43c7-bb99-0aa5476a6d83-kube-api-access-5xzs9\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-25csx\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:34 crc kubenswrapper[4912]: I1203 01:02:34.402135 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:35 crc kubenswrapper[4912]: I1203 01:02:35.087758 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx"] Dec 03 01:02:35 crc kubenswrapper[4912]: I1203 01:02:35.604815 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:02:35 crc kubenswrapper[4912]: I1203 01:02:35.976811 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" event={"ID":"afa950b8-ea20-43c7-bb99-0aa5476a6d83","Type":"ContainerStarted","Data":"9cf1018ed6c7597b191dc7ee1c7b36e89de4314ea203e76c7f54838a2f8209f6"} Dec 03 01:02:35 crc kubenswrapper[4912]: I1203 01:02:35.977376 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" event={"ID":"afa950b8-ea20-43c7-bb99-0aa5476a6d83","Type":"ContainerStarted","Data":"8c2b01b4669edbfd0cab0ebff64ad1a39ccde5bf8e68cc6c0085cf8972b6f32d"} Dec 03 01:02:36 crc kubenswrapper[4912]: I1203 01:02:36.003506 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" podStartSLOduration=1.511578895 podStartE2EDuration="2.003482029s" podCreationTimestamp="2025-12-03 01:02:34 +0000 UTC" firstStartedPulling="2025-12-03 01:02:35.105060315 +0000 UTC m=+2340.747080885" lastFinishedPulling="2025-12-03 01:02:35.596963449 +0000 UTC m=+2341.238984019" observedRunningTime="2025-12-03 01:02:35.993463611 +0000 UTC m=+2341.635484181" watchObservedRunningTime="2025-12-03 01:02:36.003482029 +0000 UTC m=+2341.645502609" Dec 03 01:02:47 crc kubenswrapper[4912]: I1203 01:02:47.097281 4912 generic.go:334] "Generic (PLEG): container finished" podID="afa950b8-ea20-43c7-bb99-0aa5476a6d83" containerID="9cf1018ed6c7597b191dc7ee1c7b36e89de4314ea203e76c7f54838a2f8209f6" exitCode=0 Dec 03 01:02:47 crc kubenswrapper[4912]: I1203 01:02:47.097414 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" event={"ID":"afa950b8-ea20-43c7-bb99-0aa5476a6d83","Type":"ContainerDied","Data":"9cf1018ed6c7597b191dc7ee1c7b36e89de4314ea203e76c7f54838a2f8209f6"} Dec 03 01:02:47 crc kubenswrapper[4912]: I1203 01:02:47.614334 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:02:47 crc kubenswrapper[4912]: E1203 01:02:47.614685 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.740110 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.742681 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-inventory\") pod \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.742732 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xzs9\" (UniqueName: \"kubernetes.io/projected/afa950b8-ea20-43c7-bb99-0aa5476a6d83-kube-api-access-5xzs9\") pod \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.742794 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-ssh-key\") pod \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\" (UID: \"afa950b8-ea20-43c7-bb99-0aa5476a6d83\") " Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.752097 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afa950b8-ea20-43c7-bb99-0aa5476a6d83-kube-api-access-5xzs9" (OuterVolumeSpecName: "kube-api-access-5xzs9") pod "afa950b8-ea20-43c7-bb99-0aa5476a6d83" (UID: "afa950b8-ea20-43c7-bb99-0aa5476a6d83"). InnerVolumeSpecName "kube-api-access-5xzs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.789755 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-inventory" (OuterVolumeSpecName: "inventory") pod "afa950b8-ea20-43c7-bb99-0aa5476a6d83" (UID: "afa950b8-ea20-43c7-bb99-0aa5476a6d83"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.813559 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "afa950b8-ea20-43c7-bb99-0aa5476a6d83" (UID: "afa950b8-ea20-43c7-bb99-0aa5476a6d83"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.845113 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xzs9\" (UniqueName: \"kubernetes.io/projected/afa950b8-ea20-43c7-bb99-0aa5476a6d83-kube-api-access-5xzs9\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.845152 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:48 crc kubenswrapper[4912]: I1203 01:02:48.845162 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/afa950b8-ea20-43c7-bb99-0aa5476a6d83-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.121686 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" event={"ID":"afa950b8-ea20-43c7-bb99-0aa5476a6d83","Type":"ContainerDied","Data":"8c2b01b4669edbfd0cab0ebff64ad1a39ccde5bf8e68cc6c0085cf8972b6f32d"} Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.122105 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c2b01b4669edbfd0cab0ebff64ad1a39ccde5bf8e68cc6c0085cf8972b6f32d" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.121771 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.215423 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr"] Dec 03 01:02:49 crc kubenswrapper[4912]: E1203 01:02:49.215979 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa950b8-ea20-43c7-bb99-0aa5476a6d83" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.216007 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa950b8-ea20-43c7-bb99-0aa5476a6d83" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.216322 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="afa950b8-ea20-43c7-bb99-0aa5476a6d83" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.217584 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.224383 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.224780 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.224949 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.226736 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.226774 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.230577 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.230731 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.230945 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.231759 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr"] Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.253401 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.253909 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.254141 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.254480 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.254631 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.254815 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.254949 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.255074 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.255265 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.255454 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.255696 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.255835 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.256037 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2fgj\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-kube-api-access-m2fgj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.357913 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.358300 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.358459 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.358641 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.358776 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.358961 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.359107 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.359266 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.359917 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.360188 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.360496 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.360666 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.360967 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2fgj\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-kube-api-access-m2fgj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.364574 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.364622 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.364630 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.365908 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.365999 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.366399 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.367269 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.367689 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.368724 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.368925 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.369337 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.370144 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.385112 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2fgj\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-kube-api-access-m2fgj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-v52jr\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:49 crc kubenswrapper[4912]: I1203 01:02:49.538050 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:02:50 crc kubenswrapper[4912]: I1203 01:02:50.156879 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr"] Dec 03 01:02:51 crc kubenswrapper[4912]: I1203 01:02:51.147332 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" event={"ID":"32607112-e615-4bb1-a05c-293dd6e96af9","Type":"ContainerStarted","Data":"e5bbe8ee1180ba8b1c51293564b93745cb0c2d026d0487f9e167c4fc3448c7c5"} Dec 03 01:02:51 crc kubenswrapper[4912]: I1203 01:02:51.147909 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" event={"ID":"32607112-e615-4bb1-a05c-293dd6e96af9","Type":"ContainerStarted","Data":"d80f133e9a8f48e9447e751d47bee53d3662081b44da2d936f079be573f98f3d"} Dec 03 01:02:51 crc kubenswrapper[4912]: I1203 01:02:51.170157 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" podStartSLOduration=1.509904462 podStartE2EDuration="2.17013399s" podCreationTimestamp="2025-12-03 01:02:49 +0000 UTC" firstStartedPulling="2025-12-03 01:02:50.160089183 +0000 UTC m=+2355.802109743" lastFinishedPulling="2025-12-03 01:02:50.820318691 +0000 UTC m=+2356.462339271" observedRunningTime="2025-12-03 01:02:51.166853982 +0000 UTC m=+2356.808874602" watchObservedRunningTime="2025-12-03 01:02:51.17013399 +0000 UTC m=+2356.812154580" Dec 03 01:02:58 crc kubenswrapper[4912]: I1203 01:02:58.572776 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:02:58 crc kubenswrapper[4912]: E1203 01:02:58.574277 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:03:10 crc kubenswrapper[4912]: I1203 01:03:10.571856 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:03:10 crc kubenswrapper[4912]: E1203 01:03:10.573163 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:03:22 crc kubenswrapper[4912]: I1203 01:03:22.572733 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:03:22 crc kubenswrapper[4912]: E1203 01:03:22.574020 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:03:33 crc kubenswrapper[4912]: I1203 01:03:33.571572 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:03:33 crc kubenswrapper[4912]: E1203 01:03:33.572911 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:03:37 crc kubenswrapper[4912]: I1203 01:03:37.723378 4912 generic.go:334] "Generic (PLEG): container finished" podID="32607112-e615-4bb1-a05c-293dd6e96af9" containerID="e5bbe8ee1180ba8b1c51293564b93745cb0c2d026d0487f9e167c4fc3448c7c5" exitCode=0 Dec 03 01:03:37 crc kubenswrapper[4912]: I1203 01:03:37.723526 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" event={"ID":"32607112-e615-4bb1-a05c-293dd6e96af9","Type":"ContainerDied","Data":"e5bbe8ee1180ba8b1c51293564b93745cb0c2d026d0487f9e167c4fc3448c7c5"} Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.287008 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.339767 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.339853 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-libvirt-combined-ca-bundle\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.339927 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340026 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-repo-setup-combined-ca-bundle\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340106 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-inventory\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340140 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-combined-ca-bundle\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340237 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340308 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ssh-key\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340344 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-bootstrap-combined-ca-bundle\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340390 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340419 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-power-monitoring-combined-ca-bundle\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340488 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2fgj\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-kube-api-access-m2fgj\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.340530 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ovn-combined-ca-bundle\") pod \"32607112-e615-4bb1-a05c-293dd6e96af9\" (UID: \"32607112-e615-4bb1-a05c-293dd6e96af9\") " Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.350083 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.350358 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.350465 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.351102 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.351503 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.353180 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.353836 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.354536 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.355147 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.357217 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-kube-api-access-m2fgj" (OuterVolumeSpecName: "kube-api-access-m2fgj") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "kube-api-access-m2fgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.357457 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.389850 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-inventory" (OuterVolumeSpecName: "inventory") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.399018 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "32607112-e615-4bb1-a05c-293dd6e96af9" (UID: "32607112-e615-4bb1-a05c-293dd6e96af9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444277 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444319 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444336 4912 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444351 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444468 4912 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444484 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2fgj\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-kube-api-access-m2fgj\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444496 4912 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444510 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444527 4912 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444549 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32607112-e615-4bb1-a05c-293dd6e96af9-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444562 4912 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444575 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.444590 4912 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32607112-e615-4bb1-a05c-293dd6e96af9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.752638 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" event={"ID":"32607112-e615-4bb1-a05c-293dd6e96af9","Type":"ContainerDied","Data":"d80f133e9a8f48e9447e751d47bee53d3662081b44da2d936f079be573f98f3d"} Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.752706 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d80f133e9a8f48e9447e751d47bee53d3662081b44da2d936f079be573f98f3d" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.752804 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.891362 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7"] Dec 03 01:03:39 crc kubenswrapper[4912]: E1203 01:03:39.891871 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32607112-e615-4bb1-a05c-293dd6e96af9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.891888 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="32607112-e615-4bb1-a05c-293dd6e96af9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.892089 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="32607112-e615-4bb1-a05c-293dd6e96af9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.893070 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.898754 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.898942 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.899461 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.900851 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.920855 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.927254 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7"] Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.961701 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9lw7\" (UniqueName: \"kubernetes.io/projected/f94ef6b3-1880-40e9-846b-30c26fd34530-kube-api-access-x9lw7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.961930 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.961976 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.962002 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f94ef6b3-1880-40e9-846b-30c26fd34530-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:39 crc kubenswrapper[4912]: I1203 01:03:39.962650 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.064566 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.064893 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.065292 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f94ef6b3-1880-40e9-846b-30c26fd34530-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.065496 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.065604 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9lw7\" (UniqueName: \"kubernetes.io/projected/f94ef6b3-1880-40e9-846b-30c26fd34530-kube-api-access-x9lw7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.066796 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f94ef6b3-1880-40e9-846b-30c26fd34530-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.069376 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.071219 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.079275 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.091976 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9lw7\" (UniqueName: \"kubernetes.io/projected/f94ef6b3-1880-40e9-846b-30c26fd34530-kube-api-access-x9lw7\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pq6h7\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.247924 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.821397 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7"] Dec 03 01:03:40 crc kubenswrapper[4912]: I1203 01:03:40.826629 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:03:41 crc kubenswrapper[4912]: I1203 01:03:41.777884 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" event={"ID":"f94ef6b3-1880-40e9-846b-30c26fd34530","Type":"ContainerStarted","Data":"4339f2e38d108b3d74d0ad6d837c554ef8af9c7ba3672f9416c09bb65e27316c"} Dec 03 01:03:42 crc kubenswrapper[4912]: I1203 01:03:42.793128 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" event={"ID":"f94ef6b3-1880-40e9-846b-30c26fd34530","Type":"ContainerStarted","Data":"bea415e1b94c3e8534e5b60c11e684118664e29409237c7273fd3e5efadebad0"} Dec 03 01:03:42 crc kubenswrapper[4912]: I1203 01:03:42.823199 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" podStartSLOduration=3.219101939 podStartE2EDuration="3.823168341s" podCreationTimestamp="2025-12-03 01:03:39 +0000 UTC" firstStartedPulling="2025-12-03 01:03:40.826275554 +0000 UTC m=+2406.468296114" lastFinishedPulling="2025-12-03 01:03:41.430341946 +0000 UTC m=+2407.072362516" observedRunningTime="2025-12-03 01:03:42.815047185 +0000 UTC m=+2408.457067835" watchObservedRunningTime="2025-12-03 01:03:42.823168341 +0000 UTC m=+2408.465188941" Dec 03 01:03:47 crc kubenswrapper[4912]: I1203 01:03:47.573633 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:03:47 crc kubenswrapper[4912]: E1203 01:03:47.574949 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:03:59 crc kubenswrapper[4912]: I1203 01:03:59.573972 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:03:59 crc kubenswrapper[4912]: E1203 01:03:59.575093 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:04:05 crc kubenswrapper[4912]: I1203 01:04:05.065402 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-69tfm"] Dec 03 01:04:05 crc kubenswrapper[4912]: I1203 01:04:05.080866 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-69tfm"] Dec 03 01:04:06 crc kubenswrapper[4912]: I1203 01:04:06.586670 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11ec043b-b8cb-4971-8fe7-041c85066c43" path="/var/lib/kubelet/pods/11ec043b-b8cb-4971-8fe7-041c85066c43/volumes" Dec 03 01:04:11 crc kubenswrapper[4912]: I1203 01:04:11.572089 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:04:11 crc kubenswrapper[4912]: E1203 01:04:11.572895 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:04:22 crc kubenswrapper[4912]: I1203 01:04:22.573025 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:04:22 crc kubenswrapper[4912]: E1203 01:04:22.574824 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:04:22 crc kubenswrapper[4912]: I1203 01:04:22.609601 4912 scope.go:117] "RemoveContainer" containerID="fc47f98c63e78123d58b819634e0d1f2249d374c9c9eedd5ff2f3a7919b47a57" Dec 03 01:04:22 crc kubenswrapper[4912]: I1203 01:04:22.652868 4912 scope.go:117] "RemoveContainer" containerID="03de48ee95541ad02f95324b6119e2e86c6a1ee323173a6bb0d2ae0a3499fc5d" Dec 03 01:04:22 crc kubenswrapper[4912]: I1203 01:04:22.729031 4912 scope.go:117] "RemoveContainer" containerID="e9a00150216afbd49e7736dc7059e1aad530ecf2a5cef18766def1ffdce4ce64" Dec 03 01:04:22 crc kubenswrapper[4912]: I1203 01:04:22.764109 4912 scope.go:117] "RemoveContainer" containerID="4d52b5310ccca02aabc692064a274fb3bcb921f03c3862b175c51b7d555080ce" Dec 03 01:04:35 crc kubenswrapper[4912]: I1203 01:04:35.572767 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:04:35 crc kubenswrapper[4912]: E1203 01:04:35.573927 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:04:49 crc kubenswrapper[4912]: I1203 01:04:49.072511 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-lfbmg"] Dec 03 01:04:49 crc kubenswrapper[4912]: I1203 01:04:49.081160 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-lfbmg"] Dec 03 01:04:50 crc kubenswrapper[4912]: I1203 01:04:50.571735 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:04:50 crc kubenswrapper[4912]: E1203 01:04:50.572355 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:04:50 crc kubenswrapper[4912]: I1203 01:04:50.584734 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de6def0-35f9-411e-9d55-475a18686b2c" path="/var/lib/kubelet/pods/1de6def0-35f9-411e-9d55-475a18686b2c/volumes" Dec 03 01:05:03 crc kubenswrapper[4912]: I1203 01:05:03.572713 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:05:03 crc kubenswrapper[4912]: E1203 01:05:03.573989 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:05:04 crc kubenswrapper[4912]: I1203 01:05:04.852789 4912 generic.go:334] "Generic (PLEG): container finished" podID="f94ef6b3-1880-40e9-846b-30c26fd34530" containerID="bea415e1b94c3e8534e5b60c11e684118664e29409237c7273fd3e5efadebad0" exitCode=0 Dec 03 01:05:04 crc kubenswrapper[4912]: I1203 01:05:04.852916 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" event={"ID":"f94ef6b3-1880-40e9-846b-30c26fd34530","Type":"ContainerDied","Data":"bea415e1b94c3e8534e5b60c11e684118664e29409237c7273fd3e5efadebad0"} Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.477080 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.519075 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9lw7\" (UniqueName: \"kubernetes.io/projected/f94ef6b3-1880-40e9-846b-30c26fd34530-kube-api-access-x9lw7\") pod \"f94ef6b3-1880-40e9-846b-30c26fd34530\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.519342 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f94ef6b3-1880-40e9-846b-30c26fd34530-ovncontroller-config-0\") pod \"f94ef6b3-1880-40e9-846b-30c26fd34530\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.519505 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-inventory\") pod \"f94ef6b3-1880-40e9-846b-30c26fd34530\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.519692 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ovn-combined-ca-bundle\") pod \"f94ef6b3-1880-40e9-846b-30c26fd34530\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.519752 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ssh-key\") pod \"f94ef6b3-1880-40e9-846b-30c26fd34530\" (UID: \"f94ef6b3-1880-40e9-846b-30c26fd34530\") " Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.537636 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f94ef6b3-1880-40e9-846b-30c26fd34530" (UID: "f94ef6b3-1880-40e9-846b-30c26fd34530"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.543540 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f94ef6b3-1880-40e9-846b-30c26fd34530-kube-api-access-x9lw7" (OuterVolumeSpecName: "kube-api-access-x9lw7") pod "f94ef6b3-1880-40e9-846b-30c26fd34530" (UID: "f94ef6b3-1880-40e9-846b-30c26fd34530"). InnerVolumeSpecName "kube-api-access-x9lw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.558345 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-inventory" (OuterVolumeSpecName: "inventory") pod "f94ef6b3-1880-40e9-846b-30c26fd34530" (UID: "f94ef6b3-1880-40e9-846b-30c26fd34530"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.573915 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f94ef6b3-1880-40e9-846b-30c26fd34530-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "f94ef6b3-1880-40e9-846b-30c26fd34530" (UID: "f94ef6b3-1880-40e9-846b-30c26fd34530"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.586711 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f94ef6b3-1880-40e9-846b-30c26fd34530" (UID: "f94ef6b3-1880-40e9-846b-30c26fd34530"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.625543 4912 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.625578 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.625589 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9lw7\" (UniqueName: \"kubernetes.io/projected/f94ef6b3-1880-40e9-846b-30c26fd34530-kube-api-access-x9lw7\") on node \"crc\" DevicePath \"\"" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.625621 4912 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/f94ef6b3-1880-40e9-846b-30c26fd34530-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.625632 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f94ef6b3-1880-40e9-846b-30c26fd34530-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.880454 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" event={"ID":"f94ef6b3-1880-40e9-846b-30c26fd34530","Type":"ContainerDied","Data":"4339f2e38d108b3d74d0ad6d837c554ef8af9c7ba3672f9416c09bb65e27316c"} Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.880514 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4339f2e38d108b3d74d0ad6d837c554ef8af9c7ba3672f9416c09bb65e27316c" Dec 03 01:05:06 crc kubenswrapper[4912]: I1203 01:05:06.880568 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.054687 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv"] Dec 03 01:05:07 crc kubenswrapper[4912]: E1203 01:05:07.055264 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f94ef6b3-1880-40e9-846b-30c26fd34530" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.055285 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f94ef6b3-1880-40e9-846b-30c26fd34530" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.055569 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f94ef6b3-1880-40e9-846b-30c26fd34530" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.056555 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.061180 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.061189 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.061531 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.061684 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.061876 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.071288 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv"] Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.139806 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.139846 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.140051 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.140076 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqqtx\" (UniqueName: \"kubernetes.io/projected/819b9191-c164-4fcd-8cf5-ed2264eaae1b-kube-api-access-jqqtx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.140107 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.242925 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.243311 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqqtx\" (UniqueName: \"kubernetes.io/projected/819b9191-c164-4fcd-8cf5-ed2264eaae1b-kube-api-access-jqqtx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.243349 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.243386 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.243408 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.249734 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.249897 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.249912 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.250139 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.261155 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqqtx\" (UniqueName: \"kubernetes.io/projected/819b9191-c164-4fcd-8cf5-ed2264eaae1b-kube-api-access-jqqtx\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dklmv\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:07 crc kubenswrapper[4912]: I1203 01:05:07.427412 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:05:08 crc kubenswrapper[4912]: I1203 01:05:08.068921 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv"] Dec 03 01:05:08 crc kubenswrapper[4912]: I1203 01:05:08.900497 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" event={"ID":"819b9191-c164-4fcd-8cf5-ed2264eaae1b","Type":"ContainerStarted","Data":"5b747f9ea44eba8564cf67b3076bf729a2da2aa8cd11f5f7fbacaedebb0a2afc"} Dec 03 01:05:09 crc kubenswrapper[4912]: I1203 01:05:09.917299 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" event={"ID":"819b9191-c164-4fcd-8cf5-ed2264eaae1b","Type":"ContainerStarted","Data":"cfc8a87e1b7079bf7027d26e0a73fee81bb31d89e7b23ffbdf5ead97447e2c4e"} Dec 03 01:05:09 crc kubenswrapper[4912]: I1203 01:05:09.938928 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" podStartSLOduration=2.327510881 podStartE2EDuration="2.938906027s" podCreationTimestamp="2025-12-03 01:05:07 +0000 UTC" firstStartedPulling="2025-12-03 01:05:08.072133086 +0000 UTC m=+2493.714153646" lastFinishedPulling="2025-12-03 01:05:08.683528222 +0000 UTC m=+2494.325548792" observedRunningTime="2025-12-03 01:05:09.937443327 +0000 UTC m=+2495.579463917" watchObservedRunningTime="2025-12-03 01:05:09.938906027 +0000 UTC m=+2495.580926597" Dec 03 01:05:15 crc kubenswrapper[4912]: I1203 01:05:15.573133 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:05:15 crc kubenswrapper[4912]: E1203 01:05:15.574758 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:05:22 crc kubenswrapper[4912]: I1203 01:05:22.876670 4912 scope.go:117] "RemoveContainer" containerID="04e8915b53cbbc4ff07d2af13a228d43ca0690a32e050c805a1efe4c43f6540f" Dec 03 01:05:27 crc kubenswrapper[4912]: I1203 01:05:27.572556 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:05:27 crc kubenswrapper[4912]: E1203 01:05:27.573501 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:05:40 crc kubenswrapper[4912]: I1203 01:05:40.572686 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:05:40 crc kubenswrapper[4912]: E1203 01:05:40.573980 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:05:51 crc kubenswrapper[4912]: I1203 01:05:51.572512 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:05:51 crc kubenswrapper[4912]: E1203 01:05:51.573543 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:06:06 crc kubenswrapper[4912]: I1203 01:06:06.574944 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:06:06 crc kubenswrapper[4912]: E1203 01:06:06.576022 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:06:20 crc kubenswrapper[4912]: I1203 01:06:20.572838 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:06:20 crc kubenswrapper[4912]: I1203 01:06:20.967637 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"4eea12316c9268b16694f56d0430dddc64cb5b7bca139041f112bb5625573ac5"} Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.327885 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ksfmg"] Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.332884 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.344889 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ksfmg"] Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.480631 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-catalog-content\") pod \"community-operators-ksfmg\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.480693 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnzs2\" (UniqueName: \"kubernetes.io/projected/1837c5c4-2107-4eb1-b905-c38eee7478e5-kube-api-access-qnzs2\") pod \"community-operators-ksfmg\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.480739 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-utilities\") pod \"community-operators-ksfmg\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.583562 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-utilities\") pod \"community-operators-ksfmg\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.583792 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-catalog-content\") pod \"community-operators-ksfmg\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.583820 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnzs2\" (UniqueName: \"kubernetes.io/projected/1837c5c4-2107-4eb1-b905-c38eee7478e5-kube-api-access-qnzs2\") pod \"community-operators-ksfmg\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.584322 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-utilities\") pod \"community-operators-ksfmg\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.584325 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-catalog-content\") pod \"community-operators-ksfmg\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.623218 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnzs2\" (UniqueName: \"kubernetes.io/projected/1837c5c4-2107-4eb1-b905-c38eee7478e5-kube-api-access-qnzs2\") pod \"community-operators-ksfmg\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:52 crc kubenswrapper[4912]: I1203 01:07:52.707540 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:07:53 crc kubenswrapper[4912]: I1203 01:07:53.325341 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ksfmg"] Dec 03 01:07:53 crc kubenswrapper[4912]: I1203 01:07:53.632276 4912 generic.go:334] "Generic (PLEG): container finished" podID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerID="72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29" exitCode=0 Dec 03 01:07:53 crc kubenswrapper[4912]: I1203 01:07:53.632358 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksfmg" event={"ID":"1837c5c4-2107-4eb1-b905-c38eee7478e5","Type":"ContainerDied","Data":"72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29"} Dec 03 01:07:53 crc kubenswrapper[4912]: I1203 01:07:53.632910 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksfmg" event={"ID":"1837c5c4-2107-4eb1-b905-c38eee7478e5","Type":"ContainerStarted","Data":"34f738c05e248f29b7afdcf8fc205eede47ee9bd40dfdca5fb10231c0c414413"} Dec 03 01:07:54 crc kubenswrapper[4912]: I1203 01:07:54.649651 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksfmg" event={"ID":"1837c5c4-2107-4eb1-b905-c38eee7478e5","Type":"ContainerStarted","Data":"1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed"} Dec 03 01:07:55 crc kubenswrapper[4912]: I1203 01:07:55.662357 4912 generic.go:334] "Generic (PLEG): container finished" podID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerID="1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed" exitCode=0 Dec 03 01:07:55 crc kubenswrapper[4912]: I1203 01:07:55.662399 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksfmg" event={"ID":"1837c5c4-2107-4eb1-b905-c38eee7478e5","Type":"ContainerDied","Data":"1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed"} Dec 03 01:07:56 crc kubenswrapper[4912]: I1203 01:07:56.682278 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksfmg" event={"ID":"1837c5c4-2107-4eb1-b905-c38eee7478e5","Type":"ContainerStarted","Data":"35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241"} Dec 03 01:07:56 crc kubenswrapper[4912]: I1203 01:07:56.725658 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ksfmg" podStartSLOduration=2.275154357 podStartE2EDuration="4.725640392s" podCreationTimestamp="2025-12-03 01:07:52 +0000 UTC" firstStartedPulling="2025-12-03 01:07:53.635615927 +0000 UTC m=+2659.277636487" lastFinishedPulling="2025-12-03 01:07:56.086101932 +0000 UTC m=+2661.728122522" observedRunningTime="2025-12-03 01:07:56.708086908 +0000 UTC m=+2662.350107468" watchObservedRunningTime="2025-12-03 01:07:56.725640392 +0000 UTC m=+2662.367660952" Dec 03 01:08:02 crc kubenswrapper[4912]: I1203 01:08:02.708870 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:08:02 crc kubenswrapper[4912]: I1203 01:08:02.709300 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:08:02 crc kubenswrapper[4912]: I1203 01:08:02.761731 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:08:02 crc kubenswrapper[4912]: I1203 01:08:02.864892 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:08:03 crc kubenswrapper[4912]: I1203 01:08:03.014336 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ksfmg"] Dec 03 01:08:04 crc kubenswrapper[4912]: I1203 01:08:04.770718 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ksfmg" podUID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerName="registry-server" containerID="cri-o://35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241" gracePeriod=2 Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.273932 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.387323 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnzs2\" (UniqueName: \"kubernetes.io/projected/1837c5c4-2107-4eb1-b905-c38eee7478e5-kube-api-access-qnzs2\") pod \"1837c5c4-2107-4eb1-b905-c38eee7478e5\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.387461 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-utilities\") pod \"1837c5c4-2107-4eb1-b905-c38eee7478e5\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.387490 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-catalog-content\") pod \"1837c5c4-2107-4eb1-b905-c38eee7478e5\" (UID: \"1837c5c4-2107-4eb1-b905-c38eee7478e5\") " Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.388798 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-utilities" (OuterVolumeSpecName: "utilities") pod "1837c5c4-2107-4eb1-b905-c38eee7478e5" (UID: "1837c5c4-2107-4eb1-b905-c38eee7478e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.402092 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1837c5c4-2107-4eb1-b905-c38eee7478e5-kube-api-access-qnzs2" (OuterVolumeSpecName: "kube-api-access-qnzs2") pod "1837c5c4-2107-4eb1-b905-c38eee7478e5" (UID: "1837c5c4-2107-4eb1-b905-c38eee7478e5"). InnerVolumeSpecName "kube-api-access-qnzs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.453576 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1837c5c4-2107-4eb1-b905-c38eee7478e5" (UID: "1837c5c4-2107-4eb1-b905-c38eee7478e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.489905 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnzs2\" (UniqueName: \"kubernetes.io/projected/1837c5c4-2107-4eb1-b905-c38eee7478e5-kube-api-access-qnzs2\") on node \"crc\" DevicePath \"\"" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.489935 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.489947 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1837c5c4-2107-4eb1-b905-c38eee7478e5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.783913 4912 generic.go:334] "Generic (PLEG): container finished" podID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerID="35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241" exitCode=0 Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.784170 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksfmg" event={"ID":"1837c5c4-2107-4eb1-b905-c38eee7478e5","Type":"ContainerDied","Data":"35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241"} Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.784525 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ksfmg" event={"ID":"1837c5c4-2107-4eb1-b905-c38eee7478e5","Type":"ContainerDied","Data":"34f738c05e248f29b7afdcf8fc205eede47ee9bd40dfdca5fb10231c0c414413"} Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.784294 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ksfmg" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.784597 4912 scope.go:117] "RemoveContainer" containerID="35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.818109 4912 scope.go:117] "RemoveContainer" containerID="1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.825811 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ksfmg"] Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.837832 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ksfmg"] Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.846343 4912 scope.go:117] "RemoveContainer" containerID="72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.914113 4912 scope.go:117] "RemoveContainer" containerID="35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241" Dec 03 01:08:05 crc kubenswrapper[4912]: E1203 01:08:05.915478 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241\": container with ID starting with 35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241 not found: ID does not exist" containerID="35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.915549 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241"} err="failed to get container status \"35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241\": rpc error: code = NotFound desc = could not find container \"35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241\": container with ID starting with 35abc3ef2c525c7026bdc5d3e83b2d49043be91eb3b8561eda62ca88d5958241 not found: ID does not exist" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.915598 4912 scope.go:117] "RemoveContainer" containerID="1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed" Dec 03 01:08:05 crc kubenswrapper[4912]: E1203 01:08:05.916078 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed\": container with ID starting with 1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed not found: ID does not exist" containerID="1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.916120 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed"} err="failed to get container status \"1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed\": rpc error: code = NotFound desc = could not find container \"1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed\": container with ID starting with 1b5c7d4932b13d3096771dbed451ca67ba0fed638320ec0b8e64078630e695ed not found: ID does not exist" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.916156 4912 scope.go:117] "RemoveContainer" containerID="72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29" Dec 03 01:08:05 crc kubenswrapper[4912]: E1203 01:08:05.916793 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29\": container with ID starting with 72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29 not found: ID does not exist" containerID="72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29" Dec 03 01:08:05 crc kubenswrapper[4912]: I1203 01:08:05.916821 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29"} err="failed to get container status \"72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29\": rpc error: code = NotFound desc = could not find container \"72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29\": container with ID starting with 72535a5740b51a4565172a091cf2af108c1c11c4a893069d9eeb41bb3220db29 not found: ID does not exist" Dec 03 01:08:06 crc kubenswrapper[4912]: I1203 01:08:06.584674 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1837c5c4-2107-4eb1-b905-c38eee7478e5" path="/var/lib/kubelet/pods/1837c5c4-2107-4eb1-b905-c38eee7478e5/volumes" Dec 03 01:08:48 crc kubenswrapper[4912]: I1203 01:08:48.078014 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:08:48 crc kubenswrapper[4912]: I1203 01:08:48.079133 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:09:18 crc kubenswrapper[4912]: I1203 01:09:18.077913 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:09:18 crc kubenswrapper[4912]: I1203 01:09:18.078554 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:09:48 crc kubenswrapper[4912]: I1203 01:09:48.077825 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:09:48 crc kubenswrapper[4912]: I1203 01:09:48.078760 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:09:48 crc kubenswrapper[4912]: I1203 01:09:48.078835 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:09:48 crc kubenswrapper[4912]: I1203 01:09:48.079864 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4eea12316c9268b16694f56d0430dddc64cb5b7bca139041f112bb5625573ac5"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:09:48 crc kubenswrapper[4912]: I1203 01:09:48.079930 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://4eea12316c9268b16694f56d0430dddc64cb5b7bca139041f112bb5625573ac5" gracePeriod=600 Dec 03 01:09:49 crc kubenswrapper[4912]: I1203 01:09:49.060805 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="4eea12316c9268b16694f56d0430dddc64cb5b7bca139041f112bb5625573ac5" exitCode=0 Dec 03 01:09:49 crc kubenswrapper[4912]: I1203 01:09:49.060863 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"4eea12316c9268b16694f56d0430dddc64cb5b7bca139041f112bb5625573ac5"} Dec 03 01:09:49 crc kubenswrapper[4912]: I1203 01:09:49.061655 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174"} Dec 03 01:09:49 crc kubenswrapper[4912]: I1203 01:09:49.061682 4912 scope.go:117] "RemoveContainer" containerID="76338a7a53c15ff37cf4f2d10acfa3af33851b524d056ecb23951313e9662ffc" Dec 03 01:10:05 crc kubenswrapper[4912]: I1203 01:10:05.291664 4912 generic.go:334] "Generic (PLEG): container finished" podID="819b9191-c164-4fcd-8cf5-ed2264eaae1b" containerID="cfc8a87e1b7079bf7027d26e0a73fee81bb31d89e7b23ffbdf5ead97447e2c4e" exitCode=0 Dec 03 01:10:05 crc kubenswrapper[4912]: I1203 01:10:05.291886 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" event={"ID":"819b9191-c164-4fcd-8cf5-ed2264eaae1b","Type":"ContainerDied","Data":"cfc8a87e1b7079bf7027d26e0a73fee81bb31d89e7b23ffbdf5ead97447e2c4e"} Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.847218 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.919618 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-secret-0\") pod \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.919958 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-ssh-key\") pod \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.920002 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-combined-ca-bundle\") pod \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.920122 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqqtx\" (UniqueName: \"kubernetes.io/projected/819b9191-c164-4fcd-8cf5-ed2264eaae1b-kube-api-access-jqqtx\") pod \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.920237 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-inventory\") pod \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\" (UID: \"819b9191-c164-4fcd-8cf5-ed2264eaae1b\") " Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.926297 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/819b9191-c164-4fcd-8cf5-ed2264eaae1b-kube-api-access-jqqtx" (OuterVolumeSpecName: "kube-api-access-jqqtx") pod "819b9191-c164-4fcd-8cf5-ed2264eaae1b" (UID: "819b9191-c164-4fcd-8cf5-ed2264eaae1b"). InnerVolumeSpecName "kube-api-access-jqqtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.927068 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "819b9191-c164-4fcd-8cf5-ed2264eaae1b" (UID: "819b9191-c164-4fcd-8cf5-ed2264eaae1b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.957819 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "819b9191-c164-4fcd-8cf5-ed2264eaae1b" (UID: "819b9191-c164-4fcd-8cf5-ed2264eaae1b"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.959843 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-inventory" (OuterVolumeSpecName: "inventory") pod "819b9191-c164-4fcd-8cf5-ed2264eaae1b" (UID: "819b9191-c164-4fcd-8cf5-ed2264eaae1b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:10:06 crc kubenswrapper[4912]: I1203 01:10:06.963697 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "819b9191-c164-4fcd-8cf5-ed2264eaae1b" (UID: "819b9191-c164-4fcd-8cf5-ed2264eaae1b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.023908 4912 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.023952 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqqtx\" (UniqueName: \"kubernetes.io/projected/819b9191-c164-4fcd-8cf5-ed2264eaae1b-kube-api-access-jqqtx\") on node \"crc\" DevicePath \"\"" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.023968 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.023982 4912 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.023997 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/819b9191-c164-4fcd-8cf5-ed2264eaae1b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.315968 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" event={"ID":"819b9191-c164-4fcd-8cf5-ed2264eaae1b","Type":"ContainerDied","Data":"5b747f9ea44eba8564cf67b3076bf729a2da2aa8cd11f5f7fbacaedebb0a2afc"} Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.316044 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b747f9ea44eba8564cf67b3076bf729a2da2aa8cd11f5f7fbacaedebb0a2afc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.316058 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.414214 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc"] Dec 03 01:10:07 crc kubenswrapper[4912]: E1203 01:10:07.415229 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerName="extract-content" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.415247 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerName="extract-content" Dec 03 01:10:07 crc kubenswrapper[4912]: E1203 01:10:07.415273 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerName="extract-utilities" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.415282 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerName="extract-utilities" Dec 03 01:10:07 crc kubenswrapper[4912]: E1203 01:10:07.415313 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerName="registry-server" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.415323 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerName="registry-server" Dec 03 01:10:07 crc kubenswrapper[4912]: E1203 01:10:07.415360 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819b9191-c164-4fcd-8cf5-ed2264eaae1b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.415369 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="819b9191-c164-4fcd-8cf5-ed2264eaae1b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.415650 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="1837c5c4-2107-4eb1-b905-c38eee7478e5" containerName="registry-server" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.415669 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="819b9191-c164-4fcd-8cf5-ed2264eaae1b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.416760 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.419691 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.419939 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.420289 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.420470 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.420533 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.429503 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc"] Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.432584 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.432702 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.433077 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.433391 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.433465 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.433590 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.433651 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-498jp\" (UniqueName: \"kubernetes.io/projected/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-kube-api-access-498jp\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.535270 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.535317 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.535355 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.535375 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-498jp\" (UniqueName: \"kubernetes.io/projected/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-kube-api-access-498jp\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.535487 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.535523 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.535569 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.539967 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.540466 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.542852 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.542975 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.543320 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.553573 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.553734 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-498jp\" (UniqueName: \"kubernetes.io/projected/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-kube-api-access-498jp\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:07 crc kubenswrapper[4912]: I1203 01:10:07.783959 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:10:08 crc kubenswrapper[4912]: I1203 01:10:08.432347 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc"] Dec 03 01:10:08 crc kubenswrapper[4912]: I1203 01:10:08.437547 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:10:09 crc kubenswrapper[4912]: I1203 01:10:09.341423 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" event={"ID":"fb8520c0-0629-4ae2-8b9d-3d3847f4c048","Type":"ContainerStarted","Data":"303c3fac11f3c5d54968f50f704d4a10d63349222450e789008ca84a96eca627"} Dec 03 01:10:09 crc kubenswrapper[4912]: I1203 01:10:09.342207 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" event={"ID":"fb8520c0-0629-4ae2-8b9d-3d3847f4c048","Type":"ContainerStarted","Data":"69ca516192531f46c226cfa095aac1b761c0e3fb38ec9d04462fbc7a7f18a96a"} Dec 03 01:10:09 crc kubenswrapper[4912]: I1203 01:10:09.361412 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" podStartSLOduration=1.797804223 podStartE2EDuration="2.361388638s" podCreationTimestamp="2025-12-03 01:10:07 +0000 UTC" firstStartedPulling="2025-12-03 01:10:08.437337762 +0000 UTC m=+2794.079358312" lastFinishedPulling="2025-12-03 01:10:09.000922157 +0000 UTC m=+2794.642942727" observedRunningTime="2025-12-03 01:10:09.361343166 +0000 UTC m=+2795.003363736" watchObservedRunningTime="2025-12-03 01:10:09.361388638 +0000 UTC m=+2795.003409238" Dec 03 01:11:48 crc kubenswrapper[4912]: I1203 01:11:48.078031 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:11:48 crc kubenswrapper[4912]: I1203 01:11:48.078723 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.057614 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5g66w"] Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.061233 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.092000 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g66w"] Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.126168 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-utilities\") pod \"redhat-marketplace-5g66w\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.126526 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-catalog-content\") pod \"redhat-marketplace-5g66w\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.126644 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz84s\" (UniqueName: \"kubernetes.io/projected/4c049250-485d-4350-8f1a-6adf65720df3-kube-api-access-sz84s\") pod \"redhat-marketplace-5g66w\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.228646 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-catalog-content\") pod \"redhat-marketplace-5g66w\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.228757 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz84s\" (UniqueName: \"kubernetes.io/projected/4c049250-485d-4350-8f1a-6adf65720df3-kube-api-access-sz84s\") pod \"redhat-marketplace-5g66w\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.228827 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-utilities\") pod \"redhat-marketplace-5g66w\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.229532 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-catalog-content\") pod \"redhat-marketplace-5g66w\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.229557 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-utilities\") pod \"redhat-marketplace-5g66w\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.261006 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz84s\" (UniqueName: \"kubernetes.io/projected/4c049250-485d-4350-8f1a-6adf65720df3-kube-api-access-sz84s\") pod \"redhat-marketplace-5g66w\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.425359 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:15 crc kubenswrapper[4912]: I1203 01:12:15.990484 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g66w"] Dec 03 01:12:16 crc kubenswrapper[4912]: I1203 01:12:16.165654 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g66w" event={"ID":"4c049250-485d-4350-8f1a-6adf65720df3","Type":"ContainerStarted","Data":"dfabd70596c7c2328315036ccf956d61e7d421e6582977d8ad6f49668d0591a4"} Dec 03 01:12:16 crc kubenswrapper[4912]: I1203 01:12:16.849206 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pxmmk"] Dec 03 01:12:16 crc kubenswrapper[4912]: I1203 01:12:16.886393 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:16 crc kubenswrapper[4912]: I1203 01:12:16.927942 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxmmk"] Dec 03 01:12:16 crc kubenswrapper[4912]: I1203 01:12:16.999181 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67383be-5842-4d3c-9a66-ce5c7075a721-catalog-content\") pod \"redhat-operators-pxmmk\" (UID: \"c67383be-5842-4d3c-9a66-ce5c7075a721\") " pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:16 crc kubenswrapper[4912]: I1203 01:12:16.999234 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67383be-5842-4d3c-9a66-ce5c7075a721-utilities\") pod \"redhat-operators-pxmmk\" (UID: \"c67383be-5842-4d3c-9a66-ce5c7075a721\") " pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:16 crc kubenswrapper[4912]: I1203 01:12:16.999462 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh2x6\" (UniqueName: \"kubernetes.io/projected/c67383be-5842-4d3c-9a66-ce5c7075a721-kube-api-access-lh2x6\") pod \"redhat-operators-pxmmk\" (UID: \"c67383be-5842-4d3c-9a66-ce5c7075a721\") " pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.101627 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67383be-5842-4d3c-9a66-ce5c7075a721-catalog-content\") pod \"redhat-operators-pxmmk\" (UID: \"c67383be-5842-4d3c-9a66-ce5c7075a721\") " pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.101747 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67383be-5842-4d3c-9a66-ce5c7075a721-utilities\") pod \"redhat-operators-pxmmk\" (UID: \"c67383be-5842-4d3c-9a66-ce5c7075a721\") " pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.101866 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh2x6\" (UniqueName: \"kubernetes.io/projected/c67383be-5842-4d3c-9a66-ce5c7075a721-kube-api-access-lh2x6\") pod \"redhat-operators-pxmmk\" (UID: \"c67383be-5842-4d3c-9a66-ce5c7075a721\") " pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.102367 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67383be-5842-4d3c-9a66-ce5c7075a721-catalog-content\") pod \"redhat-operators-pxmmk\" (UID: \"c67383be-5842-4d3c-9a66-ce5c7075a721\") " pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.102826 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67383be-5842-4d3c-9a66-ce5c7075a721-utilities\") pod \"redhat-operators-pxmmk\" (UID: \"c67383be-5842-4d3c-9a66-ce5c7075a721\") " pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.123021 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh2x6\" (UniqueName: \"kubernetes.io/projected/c67383be-5842-4d3c-9a66-ce5c7075a721-kube-api-access-lh2x6\") pod \"redhat-operators-pxmmk\" (UID: \"c67383be-5842-4d3c-9a66-ce5c7075a721\") " pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.178196 4912 generic.go:334] "Generic (PLEG): container finished" podID="4c049250-485d-4350-8f1a-6adf65720df3" containerID="9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b" exitCode=0 Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.178246 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g66w" event={"ID":"4c049250-485d-4350-8f1a-6adf65720df3","Type":"ContainerDied","Data":"9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b"} Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.217054 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.433697 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l82kx"] Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.439607 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.464998 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l82kx"] Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.512131 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-catalog-content\") pod \"certified-operators-l82kx\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.512182 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgwcq\" (UniqueName: \"kubernetes.io/projected/47c13aca-9f90-494c-91b4-e910f8e04b53-kube-api-access-rgwcq\") pod \"certified-operators-l82kx\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.512267 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-utilities\") pod \"certified-operators-l82kx\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.613968 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-catalog-content\") pod \"certified-operators-l82kx\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.614075 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgwcq\" (UniqueName: \"kubernetes.io/projected/47c13aca-9f90-494c-91b4-e910f8e04b53-kube-api-access-rgwcq\") pod \"certified-operators-l82kx\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.614512 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-catalog-content\") pod \"certified-operators-l82kx\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.615043 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-utilities\") pod \"certified-operators-l82kx\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.616070 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-utilities\") pod \"certified-operators-l82kx\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.637258 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgwcq\" (UniqueName: \"kubernetes.io/projected/47c13aca-9f90-494c-91b4-e910f8e04b53-kube-api-access-rgwcq\") pod \"certified-operators-l82kx\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.753412 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxmmk"] Dec 03 01:12:17 crc kubenswrapper[4912]: W1203 01:12:17.755832 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc67383be_5842_4d3c_9a66_ce5c7075a721.slice/crio-ca5e2865c4174debec013c048b6938ec79de90d7b0cc3f71cf480eb148828289 WatchSource:0}: Error finding container ca5e2865c4174debec013c048b6938ec79de90d7b0cc3f71cf480eb148828289: Status 404 returned error can't find the container with id ca5e2865c4174debec013c048b6938ec79de90d7b0cc3f71cf480eb148828289 Dec 03 01:12:17 crc kubenswrapper[4912]: I1203 01:12:17.779074 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:18 crc kubenswrapper[4912]: I1203 01:12:18.078333 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:12:18 crc kubenswrapper[4912]: I1203 01:12:18.078762 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:12:18 crc kubenswrapper[4912]: I1203 01:12:18.195571 4912 generic.go:334] "Generic (PLEG): container finished" podID="c67383be-5842-4d3c-9a66-ce5c7075a721" containerID="05f687fcb35f67cbc50e1124fb304c6ff182b3d6069d342b601577aefcff3e12" exitCode=0 Dec 03 01:12:18 crc kubenswrapper[4912]: I1203 01:12:18.195649 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxmmk" event={"ID":"c67383be-5842-4d3c-9a66-ce5c7075a721","Type":"ContainerDied","Data":"05f687fcb35f67cbc50e1124fb304c6ff182b3d6069d342b601577aefcff3e12"} Dec 03 01:12:18 crc kubenswrapper[4912]: I1203 01:12:18.195695 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxmmk" event={"ID":"c67383be-5842-4d3c-9a66-ce5c7075a721","Type":"ContainerStarted","Data":"ca5e2865c4174debec013c048b6938ec79de90d7b0cc3f71cf480eb148828289"} Dec 03 01:12:18 crc kubenswrapper[4912]: I1203 01:12:18.368715 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l82kx"] Dec 03 01:12:19 crc kubenswrapper[4912]: I1203 01:12:19.207884 4912 generic.go:334] "Generic (PLEG): container finished" podID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerID="3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84" exitCode=0 Dec 03 01:12:19 crc kubenswrapper[4912]: I1203 01:12:19.207934 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l82kx" event={"ID":"47c13aca-9f90-494c-91b4-e910f8e04b53","Type":"ContainerDied","Data":"3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84"} Dec 03 01:12:19 crc kubenswrapper[4912]: I1203 01:12:19.208177 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l82kx" event={"ID":"47c13aca-9f90-494c-91b4-e910f8e04b53","Type":"ContainerStarted","Data":"1d4082cc12cd55aa78506e464bd83dcefc0e58c9daf0feed879f6d38e764126b"} Dec 03 01:12:19 crc kubenswrapper[4912]: I1203 01:12:19.211151 4912 generic.go:334] "Generic (PLEG): container finished" podID="4c049250-485d-4350-8f1a-6adf65720df3" containerID="230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8" exitCode=0 Dec 03 01:12:19 crc kubenswrapper[4912]: I1203 01:12:19.211185 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g66w" event={"ID":"4c049250-485d-4350-8f1a-6adf65720df3","Type":"ContainerDied","Data":"230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8"} Dec 03 01:12:20 crc kubenswrapper[4912]: I1203 01:12:20.239411 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g66w" event={"ID":"4c049250-485d-4350-8f1a-6adf65720df3","Type":"ContainerStarted","Data":"4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678"} Dec 03 01:12:20 crc kubenswrapper[4912]: I1203 01:12:20.268263 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5g66w" podStartSLOduration=2.829614113 podStartE2EDuration="5.268240482s" podCreationTimestamp="2025-12-03 01:12:15 +0000 UTC" firstStartedPulling="2025-12-03 01:12:17.180726866 +0000 UTC m=+2922.822747426" lastFinishedPulling="2025-12-03 01:12:19.619353245 +0000 UTC m=+2925.261373795" observedRunningTime="2025-12-03 01:12:20.259404188 +0000 UTC m=+2925.901424778" watchObservedRunningTime="2025-12-03 01:12:20.268240482 +0000 UTC m=+2925.910261042" Dec 03 01:12:21 crc kubenswrapper[4912]: I1203 01:12:21.280029 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l82kx" event={"ID":"47c13aca-9f90-494c-91b4-e910f8e04b53","Type":"ContainerStarted","Data":"7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975"} Dec 03 01:12:22 crc kubenswrapper[4912]: I1203 01:12:22.299904 4912 generic.go:334] "Generic (PLEG): container finished" podID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerID="7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975" exitCode=0 Dec 03 01:12:22 crc kubenswrapper[4912]: I1203 01:12:22.299955 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l82kx" event={"ID":"47c13aca-9f90-494c-91b4-e910f8e04b53","Type":"ContainerDied","Data":"7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975"} Dec 03 01:12:25 crc kubenswrapper[4912]: I1203 01:12:25.426489 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:25 crc kubenswrapper[4912]: I1203 01:12:25.426882 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:25 crc kubenswrapper[4912]: I1203 01:12:25.477855 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:26 crc kubenswrapper[4912]: I1203 01:12:26.437351 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:26 crc kubenswrapper[4912]: I1203 01:12:26.830141 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g66w"] Dec 03 01:12:28 crc kubenswrapper[4912]: I1203 01:12:28.377174 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l82kx" event={"ID":"47c13aca-9f90-494c-91b4-e910f8e04b53","Type":"ContainerStarted","Data":"8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66"} Dec 03 01:12:28 crc kubenswrapper[4912]: I1203 01:12:28.379726 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxmmk" event={"ID":"c67383be-5842-4d3c-9a66-ce5c7075a721","Type":"ContainerStarted","Data":"b13c9ac37fe2d77f4be6dc9223f8421a8bd60e18546fbc71a1a0623002adbc8e"} Dec 03 01:12:28 crc kubenswrapper[4912]: I1203 01:12:28.379922 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5g66w" podUID="4c049250-485d-4350-8f1a-6adf65720df3" containerName="registry-server" containerID="cri-o://4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678" gracePeriod=2 Dec 03 01:12:28 crc kubenswrapper[4912]: I1203 01:12:28.415476 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l82kx" podStartSLOduration=2.7746530529999998 podStartE2EDuration="11.41544043s" podCreationTimestamp="2025-12-03 01:12:17 +0000 UTC" firstStartedPulling="2025-12-03 01:12:19.215473389 +0000 UTC m=+2924.857493949" lastFinishedPulling="2025-12-03 01:12:27.856260726 +0000 UTC m=+2933.498281326" observedRunningTime="2025-12-03 01:12:28.40677436 +0000 UTC m=+2934.048794930" watchObservedRunningTime="2025-12-03 01:12:28.41544043 +0000 UTC m=+2934.057461000" Dec 03 01:12:28 crc kubenswrapper[4912]: I1203 01:12:28.936440 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.057989 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz84s\" (UniqueName: \"kubernetes.io/projected/4c049250-485d-4350-8f1a-6adf65720df3-kube-api-access-sz84s\") pod \"4c049250-485d-4350-8f1a-6adf65720df3\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.058308 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-catalog-content\") pod \"4c049250-485d-4350-8f1a-6adf65720df3\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.058554 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-utilities\") pod \"4c049250-485d-4350-8f1a-6adf65720df3\" (UID: \"4c049250-485d-4350-8f1a-6adf65720df3\") " Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.059138 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-utilities" (OuterVolumeSpecName: "utilities") pod "4c049250-485d-4350-8f1a-6adf65720df3" (UID: "4c049250-485d-4350-8f1a-6adf65720df3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.064535 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c049250-485d-4350-8f1a-6adf65720df3-kube-api-access-sz84s" (OuterVolumeSpecName: "kube-api-access-sz84s") pod "4c049250-485d-4350-8f1a-6adf65720df3" (UID: "4c049250-485d-4350-8f1a-6adf65720df3"). InnerVolumeSpecName "kube-api-access-sz84s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.088205 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c049250-485d-4350-8f1a-6adf65720df3" (UID: "4c049250-485d-4350-8f1a-6adf65720df3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.162008 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.162075 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz84s\" (UniqueName: \"kubernetes.io/projected/4c049250-485d-4350-8f1a-6adf65720df3-kube-api-access-sz84s\") on node \"crc\" DevicePath \"\"" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.162095 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c049250-485d-4350-8f1a-6adf65720df3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.391486 4912 generic.go:334] "Generic (PLEG): container finished" podID="4c049250-485d-4350-8f1a-6adf65720df3" containerID="4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678" exitCode=0 Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.391592 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g66w" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.391640 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g66w" event={"ID":"4c049250-485d-4350-8f1a-6adf65720df3","Type":"ContainerDied","Data":"4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678"} Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.391678 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g66w" event={"ID":"4c049250-485d-4350-8f1a-6adf65720df3","Type":"ContainerDied","Data":"dfabd70596c7c2328315036ccf956d61e7d421e6582977d8ad6f49668d0591a4"} Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.391706 4912 scope.go:117] "RemoveContainer" containerID="4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.416826 4912 scope.go:117] "RemoveContainer" containerID="230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.436588 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g66w"] Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.446068 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g66w"] Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.448631 4912 scope.go:117] "RemoveContainer" containerID="9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.624338 4912 scope.go:117] "RemoveContainer" containerID="4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678" Dec 03 01:12:29 crc kubenswrapper[4912]: E1203 01:12:29.624806 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678\": container with ID starting with 4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678 not found: ID does not exist" containerID="4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.624850 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678"} err="failed to get container status \"4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678\": rpc error: code = NotFound desc = could not find container \"4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678\": container with ID starting with 4e2910165d1726552f4498c83e9e194cad566bf2104ed5b159dea24ae00da678 not found: ID does not exist" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.624879 4912 scope.go:117] "RemoveContainer" containerID="230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8" Dec 03 01:12:29 crc kubenswrapper[4912]: E1203 01:12:29.625379 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8\": container with ID starting with 230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8 not found: ID does not exist" containerID="230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.625588 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8"} err="failed to get container status \"230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8\": rpc error: code = NotFound desc = could not find container \"230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8\": container with ID starting with 230dda057d83d73cb9f7db4259da88812bbe32b77d19dde54d6ba29ae160c5b8 not found: ID does not exist" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.625638 4912 scope.go:117] "RemoveContainer" containerID="9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b" Dec 03 01:12:29 crc kubenswrapper[4912]: E1203 01:12:29.626301 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b\": container with ID starting with 9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b not found: ID does not exist" containerID="9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b" Dec 03 01:12:29 crc kubenswrapper[4912]: I1203 01:12:29.626344 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b"} err="failed to get container status \"9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b\": rpc error: code = NotFound desc = could not find container \"9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b\": container with ID starting with 9d1b786a65084113a61d8ddc4ee6f80dedbdd6fc4ec9d89d57461001ffd9f89b not found: ID does not exist" Dec 03 01:12:30 crc kubenswrapper[4912]: I1203 01:12:30.595703 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c049250-485d-4350-8f1a-6adf65720df3" path="/var/lib/kubelet/pods/4c049250-485d-4350-8f1a-6adf65720df3/volumes" Dec 03 01:12:31 crc kubenswrapper[4912]: I1203 01:12:31.419618 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxmmk" event={"ID":"c67383be-5842-4d3c-9a66-ce5c7075a721","Type":"ContainerDied","Data":"b13c9ac37fe2d77f4be6dc9223f8421a8bd60e18546fbc71a1a0623002adbc8e"} Dec 03 01:12:31 crc kubenswrapper[4912]: I1203 01:12:31.419697 4912 generic.go:334] "Generic (PLEG): container finished" podID="c67383be-5842-4d3c-9a66-ce5c7075a721" containerID="b13c9ac37fe2d77f4be6dc9223f8421a8bd60e18546fbc71a1a0623002adbc8e" exitCode=0 Dec 03 01:12:32 crc kubenswrapper[4912]: I1203 01:12:32.447271 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxmmk" event={"ID":"c67383be-5842-4d3c-9a66-ce5c7075a721","Type":"ContainerStarted","Data":"be87425795b15bda4fcda249f30e4d1351d0b3e737eb9b7e96eb308a116626c2"} Dec 03 01:12:32 crc kubenswrapper[4912]: I1203 01:12:32.481284 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pxmmk" podStartSLOduration=2.731529164 podStartE2EDuration="16.48125092s" podCreationTimestamp="2025-12-03 01:12:16 +0000 UTC" firstStartedPulling="2025-12-03 01:12:18.201869629 +0000 UTC m=+2923.843890189" lastFinishedPulling="2025-12-03 01:12:31.951591335 +0000 UTC m=+2937.593611945" observedRunningTime="2025-12-03 01:12:32.464935602 +0000 UTC m=+2938.106956232" watchObservedRunningTime="2025-12-03 01:12:32.48125092 +0000 UTC m=+2938.123271510" Dec 03 01:12:37 crc kubenswrapper[4912]: I1203 01:12:37.218490 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:37 crc kubenswrapper[4912]: I1203 01:12:37.218905 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:37 crc kubenswrapper[4912]: I1203 01:12:37.780104 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:37 crc kubenswrapper[4912]: I1203 01:12:37.780504 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:37 crc kubenswrapper[4912]: I1203 01:12:37.854000 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:38 crc kubenswrapper[4912]: I1203 01:12:38.279944 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pxmmk" podUID="c67383be-5842-4d3c-9a66-ce5c7075a721" containerName="registry-server" probeResult="failure" output=< Dec 03 01:12:38 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 01:12:38 crc kubenswrapper[4912]: > Dec 03 01:12:38 crc kubenswrapper[4912]: I1203 01:12:38.588777 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:38 crc kubenswrapper[4912]: I1203 01:12:38.660794 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l82kx"] Dec 03 01:12:40 crc kubenswrapper[4912]: I1203 01:12:40.536531 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l82kx" podUID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerName="registry-server" containerID="cri-o://8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66" gracePeriod=2 Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.064896 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.157248 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-catalog-content\") pod \"47c13aca-9f90-494c-91b4-e910f8e04b53\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.157350 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-utilities\") pod \"47c13aca-9f90-494c-91b4-e910f8e04b53\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.157373 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgwcq\" (UniqueName: \"kubernetes.io/projected/47c13aca-9f90-494c-91b4-e910f8e04b53-kube-api-access-rgwcq\") pod \"47c13aca-9f90-494c-91b4-e910f8e04b53\" (UID: \"47c13aca-9f90-494c-91b4-e910f8e04b53\") " Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.158033 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-utilities" (OuterVolumeSpecName: "utilities") pod "47c13aca-9f90-494c-91b4-e910f8e04b53" (UID: "47c13aca-9f90-494c-91b4-e910f8e04b53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.170960 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47c13aca-9f90-494c-91b4-e910f8e04b53-kube-api-access-rgwcq" (OuterVolumeSpecName: "kube-api-access-rgwcq") pod "47c13aca-9f90-494c-91b4-e910f8e04b53" (UID: "47c13aca-9f90-494c-91b4-e910f8e04b53"). InnerVolumeSpecName "kube-api-access-rgwcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.228912 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47c13aca-9f90-494c-91b4-e910f8e04b53" (UID: "47c13aca-9f90-494c-91b4-e910f8e04b53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.260843 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.260888 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47c13aca-9f90-494c-91b4-e910f8e04b53-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.260905 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgwcq\" (UniqueName: \"kubernetes.io/projected/47c13aca-9f90-494c-91b4-e910f8e04b53-kube-api-access-rgwcq\") on node \"crc\" DevicePath \"\"" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.558068 4912 generic.go:334] "Generic (PLEG): container finished" podID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerID="8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66" exitCode=0 Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.558341 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l82kx" event={"ID":"47c13aca-9f90-494c-91b4-e910f8e04b53","Type":"ContainerDied","Data":"8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66"} Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.558423 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l82kx" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.560258 4912 scope.go:117] "RemoveContainer" containerID="8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.560493 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l82kx" event={"ID":"47c13aca-9f90-494c-91b4-e910f8e04b53","Type":"ContainerDied","Data":"1d4082cc12cd55aa78506e464bd83dcefc0e58c9daf0feed879f6d38e764126b"} Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.590775 4912 scope.go:117] "RemoveContainer" containerID="7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.631028 4912 scope.go:117] "RemoveContainer" containerID="3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.636421 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l82kx"] Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.649328 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l82kx"] Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.692324 4912 scope.go:117] "RemoveContainer" containerID="8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66" Dec 03 01:12:41 crc kubenswrapper[4912]: E1203 01:12:41.693040 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66\": container with ID starting with 8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66 not found: ID does not exist" containerID="8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.693084 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66"} err="failed to get container status \"8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66\": rpc error: code = NotFound desc = could not find container \"8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66\": container with ID starting with 8871f2302b1efdbcfa3a5d0fa23d56daa3c0cfba4893609cc910c3825466de66 not found: ID does not exist" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.693108 4912 scope.go:117] "RemoveContainer" containerID="7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975" Dec 03 01:12:41 crc kubenswrapper[4912]: E1203 01:12:41.693474 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975\": container with ID starting with 7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975 not found: ID does not exist" containerID="7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.693503 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975"} err="failed to get container status \"7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975\": rpc error: code = NotFound desc = could not find container \"7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975\": container with ID starting with 7f2889acc08fbe8362a8c6dbebb2515e421b6b81f1d71e890a37763bc7c27975 not found: ID does not exist" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.693523 4912 scope.go:117] "RemoveContainer" containerID="3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84" Dec 03 01:12:41 crc kubenswrapper[4912]: E1203 01:12:41.693947 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84\": container with ID starting with 3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84 not found: ID does not exist" containerID="3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84" Dec 03 01:12:41 crc kubenswrapper[4912]: I1203 01:12:41.693964 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84"} err="failed to get container status \"3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84\": rpc error: code = NotFound desc = could not find container \"3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84\": container with ID starting with 3e853c3b530131f1576e9da9017ff2589b49112117a9d5949302b8385dcf6e84 not found: ID does not exist" Dec 03 01:12:42 crc kubenswrapper[4912]: I1203 01:12:42.587297 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47c13aca-9f90-494c-91b4-e910f8e04b53" path="/var/lib/kubelet/pods/47c13aca-9f90-494c-91b4-e910f8e04b53/volumes" Dec 03 01:12:47 crc kubenswrapper[4912]: I1203 01:12:47.319273 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:47 crc kubenswrapper[4912]: I1203 01:12:47.420204 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pxmmk" Dec 03 01:12:47 crc kubenswrapper[4912]: I1203 01:12:47.917257 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxmmk"] Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.078156 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.078231 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.078298 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.079513 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.079640 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" gracePeriod=600 Dec 03 01:12:48 crc kubenswrapper[4912]: E1203 01:12:48.204829 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.289817 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5xg4j"] Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.290218 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5xg4j" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerName="registry-server" containerID="cri-o://84aad5e461f43ab08600164c58f44b2840f3f4b03ce24164fddaef981f1aac14" gracePeriod=2 Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.693726 4912 generic.go:334] "Generic (PLEG): container finished" podID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerID="84aad5e461f43ab08600164c58f44b2840f3f4b03ce24164fddaef981f1aac14" exitCode=0 Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.694244 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xg4j" event={"ID":"770b4a36-b3be-4828-9e41-3fb24fa639f5","Type":"ContainerDied","Data":"84aad5e461f43ab08600164c58f44b2840f3f4b03ce24164fddaef981f1aac14"} Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.708778 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" exitCode=0 Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.709551 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174"} Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.709623 4912 scope.go:117] "RemoveContainer" containerID="4eea12316c9268b16694f56d0430dddc64cb5b7bca139041f112bb5625573ac5" Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.710845 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:12:48 crc kubenswrapper[4912]: E1203 01:12:48.711280 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:12:48 crc kubenswrapper[4912]: I1203 01:12:48.942139 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.083708 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-utilities\") pod \"770b4a36-b3be-4828-9e41-3fb24fa639f5\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.083984 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-catalog-content\") pod \"770b4a36-b3be-4828-9e41-3fb24fa639f5\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.084072 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t89wf\" (UniqueName: \"kubernetes.io/projected/770b4a36-b3be-4828-9e41-3fb24fa639f5-kube-api-access-t89wf\") pod \"770b4a36-b3be-4828-9e41-3fb24fa639f5\" (UID: \"770b4a36-b3be-4828-9e41-3fb24fa639f5\") " Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.085361 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-utilities" (OuterVolumeSpecName: "utilities") pod "770b4a36-b3be-4828-9e41-3fb24fa639f5" (UID: "770b4a36-b3be-4828-9e41-3fb24fa639f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.094530 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770b4a36-b3be-4828-9e41-3fb24fa639f5-kube-api-access-t89wf" (OuterVolumeSpecName: "kube-api-access-t89wf") pod "770b4a36-b3be-4828-9e41-3fb24fa639f5" (UID: "770b4a36-b3be-4828-9e41-3fb24fa639f5"). InnerVolumeSpecName "kube-api-access-t89wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.186637 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t89wf\" (UniqueName: \"kubernetes.io/projected/770b4a36-b3be-4828-9e41-3fb24fa639f5-kube-api-access-t89wf\") on node \"crc\" DevicePath \"\"" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.186671 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.204401 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "770b4a36-b3be-4828-9e41-3fb24fa639f5" (UID: "770b4a36-b3be-4828-9e41-3fb24fa639f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.288626 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/770b4a36-b3be-4828-9e41-3fb24fa639f5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.721024 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5xg4j" event={"ID":"770b4a36-b3be-4828-9e41-3fb24fa639f5","Type":"ContainerDied","Data":"d55ecddf073fc89f64f801b2e84d00605670b25fb66a9db4ef9a9ea48dd94f9c"} Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.721289 4912 scope.go:117] "RemoveContainer" containerID="84aad5e461f43ab08600164c58f44b2840f3f4b03ce24164fddaef981f1aac14" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.721406 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5xg4j" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.753546 4912 scope.go:117] "RemoveContainer" containerID="59db0e41c7e0d79491b7cca8fafd497c031b3dbf72a5d23570c46a4185ee298d" Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.759507 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5xg4j"] Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.780578 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5xg4j"] Dec 03 01:12:49 crc kubenswrapper[4912]: I1203 01:12:49.792347 4912 scope.go:117] "RemoveContainer" containerID="189976446c8dec6c87e6f44369faa6313c53cbda2d104c28e46f9cffcaf6e0fe" Dec 03 01:12:50 crc kubenswrapper[4912]: I1203 01:12:50.585117 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" path="/var/lib/kubelet/pods/770b4a36-b3be-4828-9e41-3fb24fa639f5/volumes" Dec 03 01:13:00 crc kubenswrapper[4912]: I1203 01:13:00.579593 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:13:00 crc kubenswrapper[4912]: E1203 01:13:00.582797 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:13:08 crc kubenswrapper[4912]: I1203 01:13:08.989387 4912 generic.go:334] "Generic (PLEG): container finished" podID="fb8520c0-0629-4ae2-8b9d-3d3847f4c048" containerID="303c3fac11f3c5d54968f50f704d4a10d63349222450e789008ca84a96eca627" exitCode=0 Dec 03 01:13:08 crc kubenswrapper[4912]: I1203 01:13:08.989484 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" event={"ID":"fb8520c0-0629-4ae2-8b9d-3d3847f4c048","Type":"ContainerDied","Data":"303c3fac11f3c5d54968f50f704d4a10d63349222450e789008ca84a96eca627"} Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.526297 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.631599 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-0\") pod \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.632375 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-telemetry-combined-ca-bundle\") pod \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.632458 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-inventory\") pod \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.632569 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-498jp\" (UniqueName: \"kubernetes.io/projected/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-kube-api-access-498jp\") pod \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.632609 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-2\") pod \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.632701 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-1\") pod \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.632726 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ssh-key\") pod \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\" (UID: \"fb8520c0-0629-4ae2-8b9d-3d3847f4c048\") " Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.637486 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-kube-api-access-498jp" (OuterVolumeSpecName: "kube-api-access-498jp") pod "fb8520c0-0629-4ae2-8b9d-3d3847f4c048" (UID: "fb8520c0-0629-4ae2-8b9d-3d3847f4c048"). InnerVolumeSpecName "kube-api-access-498jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.640243 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "fb8520c0-0629-4ae2-8b9d-3d3847f4c048" (UID: "fb8520c0-0629-4ae2-8b9d-3d3847f4c048"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.669557 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "fb8520c0-0629-4ae2-8b9d-3d3847f4c048" (UID: "fb8520c0-0629-4ae2-8b9d-3d3847f4c048"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.669832 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "fb8520c0-0629-4ae2-8b9d-3d3847f4c048" (UID: "fb8520c0-0629-4ae2-8b9d-3d3847f4c048"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.673142 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb8520c0-0629-4ae2-8b9d-3d3847f4c048" (UID: "fb8520c0-0629-4ae2-8b9d-3d3847f4c048"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.680229 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-inventory" (OuterVolumeSpecName: "inventory") pod "fb8520c0-0629-4ae2-8b9d-3d3847f4c048" (UID: "fb8520c0-0629-4ae2-8b9d-3d3847f4c048"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.711128 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "fb8520c0-0629-4ae2-8b9d-3d3847f4c048" (UID: "fb8520c0-0629-4ae2-8b9d-3d3847f4c048"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.737753 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.737805 4912 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.737822 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.737836 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-498jp\" (UniqueName: \"kubernetes.io/projected/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-kube-api-access-498jp\") on node \"crc\" DevicePath \"\"" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.737849 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.737863 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 01:13:10 crc kubenswrapper[4912]: I1203 01:13:10.737875 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb8520c0-0629-4ae2-8b9d-3d3847f4c048-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.019787 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" event={"ID":"fb8520c0-0629-4ae2-8b9d-3d3847f4c048","Type":"ContainerDied","Data":"69ca516192531f46c226cfa095aac1b761c0e3fb38ec9d04462fbc7a7f18a96a"} Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.019863 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69ca516192531f46c226cfa095aac1b761c0e3fb38ec9d04462fbc7a7f18a96a" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.019888 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.140769 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp"] Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141313 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerName="extract-utilities" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141341 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerName="extract-utilities" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141365 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerName="registry-server" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141374 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerName="registry-server" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141391 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c049250-485d-4350-8f1a-6adf65720df3" containerName="extract-content" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141400 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c049250-485d-4350-8f1a-6adf65720df3" containerName="extract-content" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141412 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerName="extract-content" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141420 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerName="extract-content" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141463 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerName="registry-server" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141472 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerName="registry-server" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141488 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerName="extract-utilities" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141495 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerName="extract-utilities" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141516 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb8520c0-0629-4ae2-8b9d-3d3847f4c048" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141526 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb8520c0-0629-4ae2-8b9d-3d3847f4c048" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141555 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c049250-485d-4350-8f1a-6adf65720df3" containerName="extract-utilities" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141564 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c049250-485d-4350-8f1a-6adf65720df3" containerName="extract-utilities" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141580 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c049250-485d-4350-8f1a-6adf65720df3" containerName="registry-server" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141588 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c049250-485d-4350-8f1a-6adf65720df3" containerName="registry-server" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.141605 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerName="extract-content" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141612 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerName="extract-content" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141901 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb8520c0-0629-4ae2-8b9d-3d3847f4c048" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141939 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="770b4a36-b3be-4828-9e41-3fb24fa639f5" containerName="registry-server" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141956 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="47c13aca-9f90-494c-91b4-e910f8e04b53" containerName="registry-server" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.141978 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c049250-485d-4350-8f1a-6adf65720df3" containerName="registry-server" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.142914 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.146935 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.147137 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.147460 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.147886 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.148039 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.173680 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp"] Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.247127 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.247397 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcvrp\" (UniqueName: \"kubernetes.io/projected/2014fe2a-2cba-4c73-b99e-cd77854196b2-kube-api-access-zcvrp\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.247494 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.247589 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.247641 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.247862 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.248009 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.350620 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.350784 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcvrp\" (UniqueName: \"kubernetes.io/projected/2014fe2a-2cba-4c73-b99e-cd77854196b2-kube-api-access-zcvrp\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.350874 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.350981 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.351061 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.351292 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.351524 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.355228 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.355632 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.355645 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.356210 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.356752 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.358986 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.382284 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcvrp\" (UniqueName: \"kubernetes.io/projected/2014fe2a-2cba-4c73-b99e-cd77854196b2-kube-api-access-zcvrp\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.477228 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:13:11 crc kubenswrapper[4912]: I1203 01:13:11.571817 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:13:11 crc kubenswrapper[4912]: E1203 01:13:11.572102 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:13:12 crc kubenswrapper[4912]: I1203 01:13:12.107148 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp"] Dec 03 01:13:13 crc kubenswrapper[4912]: I1203 01:13:13.048810 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" event={"ID":"2014fe2a-2cba-4c73-b99e-cd77854196b2","Type":"ContainerStarted","Data":"7c124d8dc3ea3a8de90f863adbac4f38fe8cf8145b9ed6605f57f95dece85c30"} Dec 03 01:13:13 crc kubenswrapper[4912]: I1203 01:13:13.049131 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" event={"ID":"2014fe2a-2cba-4c73-b99e-cd77854196b2","Type":"ContainerStarted","Data":"614362bd6af789e59204297a3e35483dba2692dd748a679341870bf74e4b11e9"} Dec 03 01:13:13 crc kubenswrapper[4912]: I1203 01:13:13.073998 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" podStartSLOduration=1.532902427 podStartE2EDuration="2.073978408s" podCreationTimestamp="2025-12-03 01:13:11 +0000 UTC" firstStartedPulling="2025-12-03 01:13:12.122065792 +0000 UTC m=+2977.764086352" lastFinishedPulling="2025-12-03 01:13:12.663141763 +0000 UTC m=+2978.305162333" observedRunningTime="2025-12-03 01:13:13.069603212 +0000 UTC m=+2978.711623862" watchObservedRunningTime="2025-12-03 01:13:13.073978408 +0000 UTC m=+2978.715998988" Dec 03 01:13:24 crc kubenswrapper[4912]: I1203 01:13:24.581539 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:13:24 crc kubenswrapper[4912]: E1203 01:13:24.583113 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:13:37 crc kubenswrapper[4912]: I1203 01:13:37.574382 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:13:37 crc kubenswrapper[4912]: E1203 01:13:37.575763 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:13:51 crc kubenswrapper[4912]: I1203 01:13:51.573858 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:13:51 crc kubenswrapper[4912]: E1203 01:13:51.575303 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:14:03 crc kubenswrapper[4912]: I1203 01:14:03.572230 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:14:03 crc kubenswrapper[4912]: E1203 01:14:03.573220 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:14:18 crc kubenswrapper[4912]: I1203 01:14:18.572290 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:14:18 crc kubenswrapper[4912]: E1203 01:14:18.573697 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:14:31 crc kubenswrapper[4912]: I1203 01:14:31.572782 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:14:31 crc kubenswrapper[4912]: E1203 01:14:31.573831 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:14:43 crc kubenswrapper[4912]: I1203 01:14:43.571653 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:14:43 crc kubenswrapper[4912]: E1203 01:14:43.573282 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:14:56 crc kubenswrapper[4912]: I1203 01:14:56.572707 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:14:56 crc kubenswrapper[4912]: E1203 01:14:56.574116 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.171089 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq"] Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.173705 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.176401 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.183419 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.187932 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq"] Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.241514 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crrf7\" (UniqueName: \"kubernetes.io/projected/0038382a-a3be-484a-a163-e1dc6f8cfb21-kube-api-access-crrf7\") pod \"collect-profiles-29412075-sstpq\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.241609 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0038382a-a3be-484a-a163-e1dc6f8cfb21-secret-volume\") pod \"collect-profiles-29412075-sstpq\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.241732 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0038382a-a3be-484a-a163-e1dc6f8cfb21-config-volume\") pod \"collect-profiles-29412075-sstpq\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.344543 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crrf7\" (UniqueName: \"kubernetes.io/projected/0038382a-a3be-484a-a163-e1dc6f8cfb21-kube-api-access-crrf7\") pod \"collect-profiles-29412075-sstpq\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.344623 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0038382a-a3be-484a-a163-e1dc6f8cfb21-secret-volume\") pod \"collect-profiles-29412075-sstpq\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.344695 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0038382a-a3be-484a-a163-e1dc6f8cfb21-config-volume\") pod \"collect-profiles-29412075-sstpq\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.345889 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0038382a-a3be-484a-a163-e1dc6f8cfb21-config-volume\") pod \"collect-profiles-29412075-sstpq\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.360885 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0038382a-a3be-484a-a163-e1dc6f8cfb21-secret-volume\") pod \"collect-profiles-29412075-sstpq\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.368443 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crrf7\" (UniqueName: \"kubernetes.io/projected/0038382a-a3be-484a-a163-e1dc6f8cfb21-kube-api-access-crrf7\") pod \"collect-profiles-29412075-sstpq\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:00 crc kubenswrapper[4912]: I1203 01:15:00.499830 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:01 crc kubenswrapper[4912]: I1203 01:15:01.084205 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq"] Dec 03 01:15:01 crc kubenswrapper[4912]: I1203 01:15:01.416819 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" event={"ID":"0038382a-a3be-484a-a163-e1dc6f8cfb21","Type":"ContainerStarted","Data":"6dba28d35469ad6d40a21cc1257355c24547d53f6c90226cc1f0f0027ee81ff6"} Dec 03 01:15:01 crc kubenswrapper[4912]: I1203 01:15:01.417139 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" event={"ID":"0038382a-a3be-484a-a163-e1dc6f8cfb21","Type":"ContainerStarted","Data":"80be213bd57eb3cd22620a73c213cc034a92e11607a5a39b78fb6d8a3f303f46"} Dec 03 01:15:01 crc kubenswrapper[4912]: I1203 01:15:01.437855 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" podStartSLOduration=1.437834626 podStartE2EDuration="1.437834626s" podCreationTimestamp="2025-12-03 01:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:15:01.434199129 +0000 UTC m=+3087.076219719" watchObservedRunningTime="2025-12-03 01:15:01.437834626 +0000 UTC m=+3087.079855206" Dec 03 01:15:02 crc kubenswrapper[4912]: I1203 01:15:02.428774 4912 generic.go:334] "Generic (PLEG): container finished" podID="0038382a-a3be-484a-a163-e1dc6f8cfb21" containerID="6dba28d35469ad6d40a21cc1257355c24547d53f6c90226cc1f0f0027ee81ff6" exitCode=0 Dec 03 01:15:02 crc kubenswrapper[4912]: I1203 01:15:02.428878 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" event={"ID":"0038382a-a3be-484a-a163-e1dc6f8cfb21","Type":"ContainerDied","Data":"6dba28d35469ad6d40a21cc1257355c24547d53f6c90226cc1f0f0027ee81ff6"} Dec 03 01:15:03 crc kubenswrapper[4912]: I1203 01:15:03.892949 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:03 crc kubenswrapper[4912]: I1203 01:15:03.948016 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crrf7\" (UniqueName: \"kubernetes.io/projected/0038382a-a3be-484a-a163-e1dc6f8cfb21-kube-api-access-crrf7\") pod \"0038382a-a3be-484a-a163-e1dc6f8cfb21\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " Dec 03 01:15:03 crc kubenswrapper[4912]: I1203 01:15:03.948150 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0038382a-a3be-484a-a163-e1dc6f8cfb21-secret-volume\") pod \"0038382a-a3be-484a-a163-e1dc6f8cfb21\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " Dec 03 01:15:03 crc kubenswrapper[4912]: I1203 01:15:03.948252 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0038382a-a3be-484a-a163-e1dc6f8cfb21-config-volume\") pod \"0038382a-a3be-484a-a163-e1dc6f8cfb21\" (UID: \"0038382a-a3be-484a-a163-e1dc6f8cfb21\") " Dec 03 01:15:03 crc kubenswrapper[4912]: I1203 01:15:03.949089 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0038382a-a3be-484a-a163-e1dc6f8cfb21-config-volume" (OuterVolumeSpecName: "config-volume") pod "0038382a-a3be-484a-a163-e1dc6f8cfb21" (UID: "0038382a-a3be-484a-a163-e1dc6f8cfb21"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:15:03 crc kubenswrapper[4912]: I1203 01:15:03.955006 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0038382a-a3be-484a-a163-e1dc6f8cfb21-kube-api-access-crrf7" (OuterVolumeSpecName: "kube-api-access-crrf7") pod "0038382a-a3be-484a-a163-e1dc6f8cfb21" (UID: "0038382a-a3be-484a-a163-e1dc6f8cfb21"). InnerVolumeSpecName "kube-api-access-crrf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:15:03 crc kubenswrapper[4912]: I1203 01:15:03.956948 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0038382a-a3be-484a-a163-e1dc6f8cfb21-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0038382a-a3be-484a-a163-e1dc6f8cfb21" (UID: "0038382a-a3be-484a-a163-e1dc6f8cfb21"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:15:04 crc kubenswrapper[4912]: I1203 01:15:04.050420 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0038382a-a3be-484a-a163-e1dc6f8cfb21-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:04 crc kubenswrapper[4912]: I1203 01:15:04.050757 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crrf7\" (UniqueName: \"kubernetes.io/projected/0038382a-a3be-484a-a163-e1dc6f8cfb21-kube-api-access-crrf7\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:04 crc kubenswrapper[4912]: I1203 01:15:04.050852 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0038382a-a3be-484a-a163-e1dc6f8cfb21-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:04 crc kubenswrapper[4912]: I1203 01:15:04.454904 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" event={"ID":"0038382a-a3be-484a-a163-e1dc6f8cfb21","Type":"ContainerDied","Data":"80be213bd57eb3cd22620a73c213cc034a92e11607a5a39b78fb6d8a3f303f46"} Dec 03 01:15:04 crc kubenswrapper[4912]: I1203 01:15:04.455184 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80be213bd57eb3cd22620a73c213cc034a92e11607a5a39b78fb6d8a3f303f46" Dec 03 01:15:04 crc kubenswrapper[4912]: I1203 01:15:04.455044 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq" Dec 03 01:15:04 crc kubenswrapper[4912]: I1203 01:15:04.557108 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554"] Dec 03 01:15:04 crc kubenswrapper[4912]: I1203 01:15:04.570282 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412030-2f554"] Dec 03 01:15:04 crc kubenswrapper[4912]: I1203 01:15:04.585723 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adfec43c-15ce-4a27-b6f2-3d75fe92d634" path="/var/lib/kubelet/pods/adfec43c-15ce-4a27-b6f2-3d75fe92d634/volumes" Dec 03 01:15:09 crc kubenswrapper[4912]: I1203 01:15:09.571544 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:15:09 crc kubenswrapper[4912]: E1203 01:15:09.572282 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:15:21 crc kubenswrapper[4912]: I1203 01:15:21.574584 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:15:21 crc kubenswrapper[4912]: E1203 01:15:21.576658 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:15:23 crc kubenswrapper[4912]: I1203 01:15:23.286368 4912 scope.go:117] "RemoveContainer" containerID="b9943dc5d01de284185db59710f6de9a9aa6a45df7ba4b4a7bc40169faf6ae47" Dec 03 01:15:34 crc kubenswrapper[4912]: I1203 01:15:34.602373 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:15:34 crc kubenswrapper[4912]: E1203 01:15:34.604486 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:15:43 crc kubenswrapper[4912]: I1203 01:15:43.978931 4912 generic.go:334] "Generic (PLEG): container finished" podID="2014fe2a-2cba-4c73-b99e-cd77854196b2" containerID="7c124d8dc3ea3a8de90f863adbac4f38fe8cf8145b9ed6605f57f95dece85c30" exitCode=0 Dec 03 01:15:43 crc kubenswrapper[4912]: I1203 01:15:43.979097 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" event={"ID":"2014fe2a-2cba-4c73-b99e-cd77854196b2","Type":"ContainerDied","Data":"7c124d8dc3ea3a8de90f863adbac4f38fe8cf8145b9ed6605f57f95dece85c30"} Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.540117 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.697462 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-inventory\") pod \"2014fe2a-2cba-4c73-b99e-cd77854196b2\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.697521 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ssh-key\") pod \"2014fe2a-2cba-4c73-b99e-cd77854196b2\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.697560 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcvrp\" (UniqueName: \"kubernetes.io/projected/2014fe2a-2cba-4c73-b99e-cd77854196b2-kube-api-access-zcvrp\") pod \"2014fe2a-2cba-4c73-b99e-cd77854196b2\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.697620 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-2\") pod \"2014fe2a-2cba-4c73-b99e-cd77854196b2\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.697695 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-telemetry-power-monitoring-combined-ca-bundle\") pod \"2014fe2a-2cba-4c73-b99e-cd77854196b2\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.697755 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-1\") pod \"2014fe2a-2cba-4c73-b99e-cd77854196b2\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.697814 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-0\") pod \"2014fe2a-2cba-4c73-b99e-cd77854196b2\" (UID: \"2014fe2a-2cba-4c73-b99e-cd77854196b2\") " Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.705949 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "2014fe2a-2cba-4c73-b99e-cd77854196b2" (UID: "2014fe2a-2cba-4c73-b99e-cd77854196b2"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.707472 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2014fe2a-2cba-4c73-b99e-cd77854196b2-kube-api-access-zcvrp" (OuterVolumeSpecName: "kube-api-access-zcvrp") pod "2014fe2a-2cba-4c73-b99e-cd77854196b2" (UID: "2014fe2a-2cba-4c73-b99e-cd77854196b2"). InnerVolumeSpecName "kube-api-access-zcvrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.731636 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "2014fe2a-2cba-4c73-b99e-cd77854196b2" (UID: "2014fe2a-2cba-4c73-b99e-cd77854196b2"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.742172 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2014fe2a-2cba-4c73-b99e-cd77854196b2" (UID: "2014fe2a-2cba-4c73-b99e-cd77854196b2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.744321 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "2014fe2a-2cba-4c73-b99e-cd77854196b2" (UID: "2014fe2a-2cba-4c73-b99e-cd77854196b2"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.763074 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "2014fe2a-2cba-4c73-b99e-cd77854196b2" (UID: "2014fe2a-2cba-4c73-b99e-cd77854196b2"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.763993 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-inventory" (OuterVolumeSpecName: "inventory") pod "2014fe2a-2cba-4c73-b99e-cd77854196b2" (UID: "2014fe2a-2cba-4c73-b99e-cd77854196b2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.800526 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.800563 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.800579 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcvrp\" (UniqueName: \"kubernetes.io/projected/2014fe2a-2cba-4c73-b99e-cd77854196b2-kube-api-access-zcvrp\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.800594 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.800609 4912 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.800624 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:45 crc kubenswrapper[4912]: I1203 01:15:45.800638 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/2014fe2a-2cba-4c73-b99e-cd77854196b2-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.005636 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" event={"ID":"2014fe2a-2cba-4c73-b99e-cd77854196b2","Type":"ContainerDied","Data":"614362bd6af789e59204297a3e35483dba2692dd748a679341870bf74e4b11e9"} Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.005683 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="614362bd6af789e59204297a3e35483dba2692dd748a679341870bf74e4b11e9" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.005720 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.144525 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx"] Dec 03 01:15:46 crc kubenswrapper[4912]: E1203 01:15:46.145641 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0038382a-a3be-484a-a163-e1dc6f8cfb21" containerName="collect-profiles" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.145679 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0038382a-a3be-484a-a163-e1dc6f8cfb21" containerName="collect-profiles" Dec 03 01:15:46 crc kubenswrapper[4912]: E1203 01:15:46.145723 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2014fe2a-2cba-4c73-b99e-cd77854196b2" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.145741 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2014fe2a-2cba-4c73-b99e-cd77854196b2" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.146155 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2014fe2a-2cba-4c73-b99e-cd77854196b2" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.146214 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0038382a-a3be-484a-a163-e1dc6f8cfb21" containerName="collect-profiles" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.147564 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.150950 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.151099 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.151099 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.151492 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.154415 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.155967 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx"] Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.311266 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwp5p\" (UniqueName: \"kubernetes.io/projected/e41b7f71-1bfd-4821-b68e-f6f015bfe072-kube-api-access-bwp5p\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.311386 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.311654 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.311914 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.312081 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.414419 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.414540 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.414595 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.414635 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.414738 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwp5p\" (UniqueName: \"kubernetes.io/projected/e41b7f71-1bfd-4821-b68e-f6f015bfe072-kube-api-access-bwp5p\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.420321 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.421032 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.421816 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.422638 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.443875 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwp5p\" (UniqueName: \"kubernetes.io/projected/e41b7f71-1bfd-4821-b68e-f6f015bfe072-kube-api-access-bwp5p\") pod \"logging-edpm-deployment-openstack-edpm-ipam-6mszx\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.518952 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:15:46 crc kubenswrapper[4912]: I1203 01:15:46.573141 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:15:46 crc kubenswrapper[4912]: E1203 01:15:46.573838 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:15:47 crc kubenswrapper[4912]: W1203 01:15:47.137318 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode41b7f71_1bfd_4821_b68e_f6f015bfe072.slice/crio-880f5847aa16e4c62cdef17c8815380a23f959da0bdade32e3de84500de8ad41 WatchSource:0}: Error finding container 880f5847aa16e4c62cdef17c8815380a23f959da0bdade32e3de84500de8ad41: Status 404 returned error can't find the container with id 880f5847aa16e4c62cdef17c8815380a23f959da0bdade32e3de84500de8ad41 Dec 03 01:15:47 crc kubenswrapper[4912]: I1203 01:15:47.142348 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:15:47 crc kubenswrapper[4912]: I1203 01:15:47.152749 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx"] Dec 03 01:15:48 crc kubenswrapper[4912]: I1203 01:15:48.033919 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" event={"ID":"e41b7f71-1bfd-4821-b68e-f6f015bfe072","Type":"ContainerStarted","Data":"f91418c182e59bd5f6225519590b349d8193d859b3d08c0a4d07864c1de304e5"} Dec 03 01:15:48 crc kubenswrapper[4912]: I1203 01:15:48.034269 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" event={"ID":"e41b7f71-1bfd-4821-b68e-f6f015bfe072","Type":"ContainerStarted","Data":"880f5847aa16e4c62cdef17c8815380a23f959da0bdade32e3de84500de8ad41"} Dec 03 01:15:48 crc kubenswrapper[4912]: I1203 01:15:48.079868 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" podStartSLOduration=1.539769255 podStartE2EDuration="2.079831845s" podCreationTimestamp="2025-12-03 01:15:46 +0000 UTC" firstStartedPulling="2025-12-03 01:15:47.141717731 +0000 UTC m=+3132.783738321" lastFinishedPulling="2025-12-03 01:15:47.681780341 +0000 UTC m=+3133.323800911" observedRunningTime="2025-12-03 01:15:48.063759944 +0000 UTC m=+3133.705780534" watchObservedRunningTime="2025-12-03 01:15:48.079831845 +0000 UTC m=+3133.721852445" Dec 03 01:15:59 crc kubenswrapper[4912]: I1203 01:15:59.572115 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:15:59 crc kubenswrapper[4912]: E1203 01:15:59.572936 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:16:09 crc kubenswrapper[4912]: I1203 01:16:09.291102 4912 generic.go:334] "Generic (PLEG): container finished" podID="e41b7f71-1bfd-4821-b68e-f6f015bfe072" containerID="f91418c182e59bd5f6225519590b349d8193d859b3d08c0a4d07864c1de304e5" exitCode=0 Dec 03 01:16:09 crc kubenswrapper[4912]: I1203 01:16:09.291267 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" event={"ID":"e41b7f71-1bfd-4821-b68e-f6f015bfe072","Type":"ContainerDied","Data":"f91418c182e59bd5f6225519590b349d8193d859b3d08c0a4d07864c1de304e5"} Dec 03 01:16:10 crc kubenswrapper[4912]: I1203 01:16:10.870215 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.015591 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwp5p\" (UniqueName: \"kubernetes.io/projected/e41b7f71-1bfd-4821-b68e-f6f015bfe072-kube-api-access-bwp5p\") pod \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.015750 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-0\") pod \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.015786 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-inventory\") pod \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.015913 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-ssh-key\") pod \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.016035 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-1\") pod \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.027144 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e41b7f71-1bfd-4821-b68e-f6f015bfe072-kube-api-access-bwp5p" (OuterVolumeSpecName: "kube-api-access-bwp5p") pod "e41b7f71-1bfd-4821-b68e-f6f015bfe072" (UID: "e41b7f71-1bfd-4821-b68e-f6f015bfe072"). InnerVolumeSpecName "kube-api-access-bwp5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.051601 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "e41b7f71-1bfd-4821-b68e-f6f015bfe072" (UID: "e41b7f71-1bfd-4821-b68e-f6f015bfe072"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:16:11 crc kubenswrapper[4912]: E1203 01:16:11.062961 4912 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-0 podName:e41b7f71-1bfd-4821-b68e-f6f015bfe072 nodeName:}" failed. No retries permitted until 2025-12-03 01:16:11.562934682 +0000 UTC m=+3157.204955242 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "logging-compute-config-data-0" (UniqueName: "kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-0") pod "e41b7f71-1bfd-4821-b68e-f6f015bfe072" (UID: "e41b7f71-1bfd-4821-b68e-f6f015bfe072") : error deleting /var/lib/kubelet/pods/e41b7f71-1bfd-4821-b68e-f6f015bfe072/volume-subpaths: remove /var/lib/kubelet/pods/e41b7f71-1bfd-4821-b68e-f6f015bfe072/volume-subpaths: no such file or directory Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.065750 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e41b7f71-1bfd-4821-b68e-f6f015bfe072" (UID: "e41b7f71-1bfd-4821-b68e-f6f015bfe072"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.070247 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-inventory" (OuterVolumeSpecName: "inventory") pod "e41b7f71-1bfd-4821-b68e-f6f015bfe072" (UID: "e41b7f71-1bfd-4821-b68e-f6f015bfe072"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.118733 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.118765 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.118784 4912 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.118796 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwp5p\" (UniqueName: \"kubernetes.io/projected/e41b7f71-1bfd-4821-b68e-f6f015bfe072-kube-api-access-bwp5p\") on node \"crc\" DevicePath \"\"" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.314788 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" event={"ID":"e41b7f71-1bfd-4821-b68e-f6f015bfe072","Type":"ContainerDied","Data":"880f5847aa16e4c62cdef17c8815380a23f959da0bdade32e3de84500de8ad41"} Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.315178 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="880f5847aa16e4c62cdef17c8815380a23f959da0bdade32e3de84500de8ad41" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.314862 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.629836 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-0\") pod \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\" (UID: \"e41b7f71-1bfd-4821-b68e-f6f015bfe072\") " Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.636629 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "e41b7f71-1bfd-4821-b68e-f6f015bfe072" (UID: "e41b7f71-1bfd-4821-b68e-f6f015bfe072"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:16:11 crc kubenswrapper[4912]: I1203 01:16:11.733203 4912 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e41b7f71-1bfd-4821-b68e-f6f015bfe072-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:16:14 crc kubenswrapper[4912]: I1203 01:16:14.587648 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:16:14 crc kubenswrapper[4912]: E1203 01:16:14.588704 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:16:25 crc kubenswrapper[4912]: I1203 01:16:25.572315 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:16:25 crc kubenswrapper[4912]: E1203 01:16:25.573029 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:16:36 crc kubenswrapper[4912]: I1203 01:16:36.573105 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:16:36 crc kubenswrapper[4912]: E1203 01:16:36.574769 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:16:48 crc kubenswrapper[4912]: I1203 01:16:48.572746 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:16:48 crc kubenswrapper[4912]: E1203 01:16:48.573878 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:17:01 crc kubenswrapper[4912]: I1203 01:17:01.573722 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:17:01 crc kubenswrapper[4912]: E1203 01:17:01.577638 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:17:15 crc kubenswrapper[4912]: I1203 01:17:15.572154 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:17:15 crc kubenswrapper[4912]: E1203 01:17:15.573342 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:17:27 crc kubenswrapper[4912]: I1203 01:17:27.572693 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:17:27 crc kubenswrapper[4912]: E1203 01:17:27.574295 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:17:41 crc kubenswrapper[4912]: I1203 01:17:41.573033 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:17:41 crc kubenswrapper[4912]: E1203 01:17:41.574106 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:17:55 crc kubenswrapper[4912]: I1203 01:17:55.573328 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:17:56 crc kubenswrapper[4912]: I1203 01:17:56.712205 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"ea5dabf8afee7141982885ac03babc8ac23410c3b3d263c54df4ae0e2cbbdf30"} Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.556797 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dzcdw"] Dec 03 01:18:58 crc kubenswrapper[4912]: E1203 01:18:58.559031 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e41b7f71-1bfd-4821-b68e-f6f015bfe072" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.559155 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e41b7f71-1bfd-4821-b68e-f6f015bfe072" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.559551 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="e41b7f71-1bfd-4821-b68e-f6f015bfe072" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.561623 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.583714 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dzcdw"] Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.602248 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkqch\" (UniqueName: \"kubernetes.io/projected/0f6b85f3-487e-4008-8a2a-3b3434fa333c-kube-api-access-jkqch\") pod \"community-operators-dzcdw\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.602524 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-catalog-content\") pod \"community-operators-dzcdw\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.602666 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-utilities\") pod \"community-operators-dzcdw\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.705562 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkqch\" (UniqueName: \"kubernetes.io/projected/0f6b85f3-487e-4008-8a2a-3b3434fa333c-kube-api-access-jkqch\") pod \"community-operators-dzcdw\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.705812 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-catalog-content\") pod \"community-operators-dzcdw\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.706187 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-catalog-content\") pod \"community-operators-dzcdw\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.706479 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-utilities\") pod \"community-operators-dzcdw\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.706888 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-utilities\") pod \"community-operators-dzcdw\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.724972 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkqch\" (UniqueName: \"kubernetes.io/projected/0f6b85f3-487e-4008-8a2a-3b3434fa333c-kube-api-access-jkqch\") pod \"community-operators-dzcdw\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:58 crc kubenswrapper[4912]: I1203 01:18:58.886051 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:18:59 crc kubenswrapper[4912]: I1203 01:18:59.442148 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dzcdw"] Dec 03 01:18:59 crc kubenswrapper[4912]: I1203 01:18:59.487338 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzcdw" event={"ID":"0f6b85f3-487e-4008-8a2a-3b3434fa333c","Type":"ContainerStarted","Data":"29e7478afd4751c19d561c3f63d18016e97c7dce84d36aef89e730d1b79bbb08"} Dec 03 01:19:00 crc kubenswrapper[4912]: I1203 01:19:00.501988 4912 generic.go:334] "Generic (PLEG): container finished" podID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerID="52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5" exitCode=0 Dec 03 01:19:00 crc kubenswrapper[4912]: I1203 01:19:00.502039 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzcdw" event={"ID":"0f6b85f3-487e-4008-8a2a-3b3434fa333c","Type":"ContainerDied","Data":"52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5"} Dec 03 01:19:02 crc kubenswrapper[4912]: I1203 01:19:02.530187 4912 generic.go:334] "Generic (PLEG): container finished" podID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerID="ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc" exitCode=0 Dec 03 01:19:02 crc kubenswrapper[4912]: I1203 01:19:02.530301 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzcdw" event={"ID":"0f6b85f3-487e-4008-8a2a-3b3434fa333c","Type":"ContainerDied","Data":"ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc"} Dec 03 01:19:03 crc kubenswrapper[4912]: I1203 01:19:03.544037 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzcdw" event={"ID":"0f6b85f3-487e-4008-8a2a-3b3434fa333c","Type":"ContainerStarted","Data":"82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed"} Dec 03 01:19:03 crc kubenswrapper[4912]: I1203 01:19:03.569808 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dzcdw" podStartSLOduration=3.113784939 podStartE2EDuration="5.569790177s" podCreationTimestamp="2025-12-03 01:18:58 +0000 UTC" firstStartedPulling="2025-12-03 01:19:00.504861444 +0000 UTC m=+3326.146882014" lastFinishedPulling="2025-12-03 01:19:02.960866682 +0000 UTC m=+3328.602887252" observedRunningTime="2025-12-03 01:19:03.562531832 +0000 UTC m=+3329.204552412" watchObservedRunningTime="2025-12-03 01:19:03.569790177 +0000 UTC m=+3329.211810738" Dec 03 01:19:08 crc kubenswrapper[4912]: I1203 01:19:08.886577 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:19:08 crc kubenswrapper[4912]: I1203 01:19:08.888806 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:19:08 crc kubenswrapper[4912]: I1203 01:19:08.970213 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:19:09 crc kubenswrapper[4912]: I1203 01:19:09.668400 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:19:09 crc kubenswrapper[4912]: I1203 01:19:09.726070 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dzcdw"] Dec 03 01:19:11 crc kubenswrapper[4912]: I1203 01:19:11.639511 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dzcdw" podUID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerName="registry-server" containerID="cri-o://82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed" gracePeriod=2 Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.291488 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.444357 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-catalog-content\") pod \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.444588 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkqch\" (UniqueName: \"kubernetes.io/projected/0f6b85f3-487e-4008-8a2a-3b3434fa333c-kube-api-access-jkqch\") pod \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.444731 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-utilities\") pod \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\" (UID: \"0f6b85f3-487e-4008-8a2a-3b3434fa333c\") " Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.446929 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-utilities" (OuterVolumeSpecName: "utilities") pod "0f6b85f3-487e-4008-8a2a-3b3434fa333c" (UID: "0f6b85f3-487e-4008-8a2a-3b3434fa333c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.454666 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f6b85f3-487e-4008-8a2a-3b3434fa333c-kube-api-access-jkqch" (OuterVolumeSpecName: "kube-api-access-jkqch") pod "0f6b85f3-487e-4008-8a2a-3b3434fa333c" (UID: "0f6b85f3-487e-4008-8a2a-3b3434fa333c"). InnerVolumeSpecName "kube-api-access-jkqch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.496335 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f6b85f3-487e-4008-8a2a-3b3434fa333c" (UID: "0f6b85f3-487e-4008-8a2a-3b3434fa333c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.547299 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.547330 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f6b85f3-487e-4008-8a2a-3b3434fa333c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.547341 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkqch\" (UniqueName: \"kubernetes.io/projected/0f6b85f3-487e-4008-8a2a-3b3434fa333c-kube-api-access-jkqch\") on node \"crc\" DevicePath \"\"" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.657114 4912 generic.go:334] "Generic (PLEG): container finished" podID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerID="82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed" exitCode=0 Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.657205 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzcdw" event={"ID":"0f6b85f3-487e-4008-8a2a-3b3434fa333c","Type":"ContainerDied","Data":"82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed"} Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.657260 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dzcdw" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.657280 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dzcdw" event={"ID":"0f6b85f3-487e-4008-8a2a-3b3434fa333c","Type":"ContainerDied","Data":"29e7478afd4751c19d561c3f63d18016e97c7dce84d36aef89e730d1b79bbb08"} Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.657311 4912 scope.go:117] "RemoveContainer" containerID="82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.683806 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dzcdw"] Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.692036 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dzcdw"] Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.698014 4912 scope.go:117] "RemoveContainer" containerID="ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.721646 4912 scope.go:117] "RemoveContainer" containerID="52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.778679 4912 scope.go:117] "RemoveContainer" containerID="82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed" Dec 03 01:19:12 crc kubenswrapper[4912]: E1203 01:19:12.780007 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed\": container with ID starting with 82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed not found: ID does not exist" containerID="82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.780052 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed"} err="failed to get container status \"82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed\": rpc error: code = NotFound desc = could not find container \"82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed\": container with ID starting with 82627d2e3621985b58a7b0c945a473a5b9566051d05c0634d4a909c13dfa8fed not found: ID does not exist" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.780084 4912 scope.go:117] "RemoveContainer" containerID="ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc" Dec 03 01:19:12 crc kubenswrapper[4912]: E1203 01:19:12.780617 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc\": container with ID starting with ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc not found: ID does not exist" containerID="ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.780679 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc"} err="failed to get container status \"ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc\": rpc error: code = NotFound desc = could not find container \"ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc\": container with ID starting with ecc841ee9c165e2af315fd498c2f8f703131e3b6f197601b5e4729f70c58d0dc not found: ID does not exist" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.780712 4912 scope.go:117] "RemoveContainer" containerID="52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5" Dec 03 01:19:12 crc kubenswrapper[4912]: E1203 01:19:12.781041 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5\": container with ID starting with 52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5 not found: ID does not exist" containerID="52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5" Dec 03 01:19:12 crc kubenswrapper[4912]: I1203 01:19:12.781069 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5"} err="failed to get container status \"52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5\": rpc error: code = NotFound desc = could not find container \"52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5\": container with ID starting with 52fe45ecb7145e972c8b9231df2414fc81b9f4eb7b57c1293043d3c2416c64f5 not found: ID does not exist" Dec 03 01:19:14 crc kubenswrapper[4912]: I1203 01:19:14.594265 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" path="/var/lib/kubelet/pods/0f6b85f3-487e-4008-8a2a-3b3434fa333c/volumes" Dec 03 01:20:18 crc kubenswrapper[4912]: I1203 01:20:18.077834 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:20:18 crc kubenswrapper[4912]: I1203 01:20:18.079164 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:20:48 crc kubenswrapper[4912]: I1203 01:20:48.077399 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:20:48 crc kubenswrapper[4912]: I1203 01:20:48.078135 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:21:18 crc kubenswrapper[4912]: I1203 01:21:18.078344 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:21:18 crc kubenswrapper[4912]: I1203 01:21:18.079109 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:21:18 crc kubenswrapper[4912]: I1203 01:21:18.079174 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:21:18 crc kubenswrapper[4912]: I1203 01:21:18.080399 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea5dabf8afee7141982885ac03babc8ac23410c3b3d263c54df4ae0e2cbbdf30"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:21:18 crc kubenswrapper[4912]: I1203 01:21:18.080562 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://ea5dabf8afee7141982885ac03babc8ac23410c3b3d263c54df4ae0e2cbbdf30" gracePeriod=600 Dec 03 01:21:18 crc kubenswrapper[4912]: I1203 01:21:18.281416 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="ea5dabf8afee7141982885ac03babc8ac23410c3b3d263c54df4ae0e2cbbdf30" exitCode=0 Dec 03 01:21:18 crc kubenswrapper[4912]: I1203 01:21:18.282127 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"ea5dabf8afee7141982885ac03babc8ac23410c3b3d263c54df4ae0e2cbbdf30"} Dec 03 01:21:18 crc kubenswrapper[4912]: I1203 01:21:18.282279 4912 scope.go:117] "RemoveContainer" containerID="8766bacb25a9fc395187ea6c981d636b738ec5ab523c8f1aa7ea2322ceaf5174" Dec 03 01:21:19 crc kubenswrapper[4912]: I1203 01:21:19.294569 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2"} Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.241410 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.258471 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.278094 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.287529 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.297917 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.306314 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.315185 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.325055 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-29v75"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.334024 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-6mszx"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.341283 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.348706 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zmwqm"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.356762 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.363683 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.370642 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.377773 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.384871 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9vcnf"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.402790 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xmrsg"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.418794 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.428551 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.435991 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.444610 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-htl96"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.452411 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-k76xp"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.459774 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-v52jr"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.466183 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-pr7lc"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.473506 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dklmv"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.481652 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-stfcg"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.489370 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c4nwx"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.500674 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-zmwqm"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.512163 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-25csx"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.522928 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-5cngp"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.533215 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pq6h7"] Dec 03 01:21:35 crc kubenswrapper[4912]: I1203 01:21:35.543839 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62pbx"] Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.588046 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="066dd271-4382-440a-9968-c02d9393cbd1" path="/var/lib/kubelet/pods/066dd271-4382-440a-9968-c02d9393cbd1/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.589633 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0908ef60-cdec-46de-beb6-167295e04796" path="/var/lib/kubelet/pods/0908ef60-cdec-46de-beb6-167295e04796/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.590244 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c3f1295-f4b6-449f-b20e-229f2e89a32a" path="/var/lib/kubelet/pods/1c3f1295-f4b6-449f-b20e-229f2e89a32a/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.590913 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2014fe2a-2cba-4c73-b99e-cd77854196b2" path="/var/lib/kubelet/pods/2014fe2a-2cba-4c73-b99e-cd77854196b2/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.592188 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32607112-e615-4bb1-a05c-293dd6e96af9" path="/var/lib/kubelet/pods/32607112-e615-4bb1-a05c-293dd6e96af9/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.592913 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b592c0a-8523-4584-9a2e-c616ea7d4ecd" path="/var/lib/kubelet/pods/4b592c0a-8523-4584-9a2e-c616ea7d4ecd/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.593667 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78887e9b-6ce9-4fc7-92ec-3c0b26052704" path="/var/lib/kubelet/pods/78887e9b-6ce9-4fc7-92ec-3c0b26052704/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.594817 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="819b9191-c164-4fcd-8cf5-ed2264eaae1b" path="/var/lib/kubelet/pods/819b9191-c164-4fcd-8cf5-ed2264eaae1b/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.595445 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85124524-226b-46f3-b4c1-3675bfc1c15f" path="/var/lib/kubelet/pods/85124524-226b-46f3-b4c1-3675bfc1c15f/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.596103 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e6b4897-744a-4650-b924-28a515fa875f" path="/var/lib/kubelet/pods/8e6b4897-744a-4650-b924-28a515fa875f/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.597267 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afa950b8-ea20-43c7-bb99-0aa5476a6d83" path="/var/lib/kubelet/pods/afa950b8-ea20-43c7-bb99-0aa5476a6d83/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.597938 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9" path="/var/lib/kubelet/pods/c2d5a24c-8277-4577-b4c3-3eb6c9a7b9c9/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.598610 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e41b7f71-1bfd-4821-b68e-f6f015bfe072" path="/var/lib/kubelet/pods/e41b7f71-1bfd-4821-b68e-f6f015bfe072/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.599328 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef3338b7-de78-4aea-8f3b-809db2358f0e" path="/var/lib/kubelet/pods/ef3338b7-de78-4aea-8f3b-809db2358f0e/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.600649 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f94ef6b3-1880-40e9-846b-30c26fd34530" path="/var/lib/kubelet/pods/f94ef6b3-1880-40e9-846b-30c26fd34530/volumes" Dec 03 01:21:36 crc kubenswrapper[4912]: I1203 01:21:36.601303 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb8520c0-0629-4ae2-8b9d-3d3847f4c048" path="/var/lib/kubelet/pods/fb8520c0-0629-4ae2-8b9d-3d3847f4c048/volumes" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.525182 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz"] Dec 03 01:21:40 crc kubenswrapper[4912]: E1203 01:21:40.526511 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerName="extract-content" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.526531 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerName="extract-content" Dec 03 01:21:40 crc kubenswrapper[4912]: E1203 01:21:40.526543 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerName="registry-server" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.526551 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerName="registry-server" Dec 03 01:21:40 crc kubenswrapper[4912]: E1203 01:21:40.526588 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerName="extract-utilities" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.526597 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerName="extract-utilities" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.526882 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f6b85f3-487e-4008-8a2a-3b3434fa333c" containerName="registry-server" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.527755 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.536746 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.537041 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.537101 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.537219 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.537339 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.558566 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz"] Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.629301 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bkk6\" (UniqueName: \"kubernetes.io/projected/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-kube-api-access-9bkk6\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.629401 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.629467 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.629519 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.629551 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.731297 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bkk6\" (UniqueName: \"kubernetes.io/projected/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-kube-api-access-9bkk6\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.731363 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.731405 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.731474 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.731493 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.738863 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.738962 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.742971 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.743552 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.752466 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bkk6\" (UniqueName: \"kubernetes.io/projected/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-kube-api-access-9bkk6\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:40 crc kubenswrapper[4912]: I1203 01:21:40.861690 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:41 crc kubenswrapper[4912]: I1203 01:21:41.430517 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz"] Dec 03 01:21:41 crc kubenswrapper[4912]: I1203 01:21:41.432818 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:21:41 crc kubenswrapper[4912]: I1203 01:21:41.625739 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" event={"ID":"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784","Type":"ContainerStarted","Data":"2880976f86a26fde45cc04cc055f436d4213ac52086779f99653f38a84dd43dd"} Dec 03 01:21:42 crc kubenswrapper[4912]: I1203 01:21:42.637929 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" event={"ID":"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784","Type":"ContainerStarted","Data":"845fefcaa4523668330b3126d31678b119c367dc166d68bd5cad8ef6420deba2"} Dec 03 01:21:42 crc kubenswrapper[4912]: I1203 01:21:42.672044 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" podStartSLOduration=2.114977527 podStartE2EDuration="2.672026181s" podCreationTimestamp="2025-12-03 01:21:40 +0000 UTC" firstStartedPulling="2025-12-03 01:21:41.432407156 +0000 UTC m=+3487.074427746" lastFinishedPulling="2025-12-03 01:21:41.98945583 +0000 UTC m=+3487.631476400" observedRunningTime="2025-12-03 01:21:42.658677147 +0000 UTC m=+3488.300697707" watchObservedRunningTime="2025-12-03 01:21:42.672026181 +0000 UTC m=+3488.314046731" Dec 03 01:21:55 crc kubenswrapper[4912]: I1203 01:21:55.879140 4912 generic.go:334] "Generic (PLEG): container finished" podID="0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784" containerID="845fefcaa4523668330b3126d31678b119c367dc166d68bd5cad8ef6420deba2" exitCode=0 Dec 03 01:21:55 crc kubenswrapper[4912]: I1203 01:21:55.879851 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" event={"ID":"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784","Type":"ContainerDied","Data":"845fefcaa4523668330b3126d31678b119c367dc166d68bd5cad8ef6420deba2"} Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.461049 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.591260 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ceph\") pod \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.591464 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ssh-key\") pod \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.591637 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-repo-setup-combined-ca-bundle\") pod \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.591703 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-inventory\") pod \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.591805 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bkk6\" (UniqueName: \"kubernetes.io/projected/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-kube-api-access-9bkk6\") pod \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\" (UID: \"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784\") " Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.599619 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784" (UID: "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.599649 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-kube-api-access-9bkk6" (OuterVolumeSpecName: "kube-api-access-9bkk6") pod "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784" (UID: "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784"). InnerVolumeSpecName "kube-api-access-9bkk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.602128 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ceph" (OuterVolumeSpecName: "ceph") pod "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784" (UID: "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.627485 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784" (UID: "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.638411 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-inventory" (OuterVolumeSpecName: "inventory") pod "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784" (UID: "0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.695777 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bkk6\" (UniqueName: \"kubernetes.io/projected/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-kube-api-access-9bkk6\") on node \"crc\" DevicePath \"\"" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.695819 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.695835 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.695847 4912 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.695861 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.912144 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" event={"ID":"0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784","Type":"ContainerDied","Data":"2880976f86a26fde45cc04cc055f436d4213ac52086779f99653f38a84dd43dd"} Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.912195 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2880976f86a26fde45cc04cc055f436d4213ac52086779f99653f38a84dd43dd" Dec 03 01:21:57 crc kubenswrapper[4912]: I1203 01:21:57.912265 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.010340 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr"] Dec 03 01:21:58 crc kubenswrapper[4912]: E1203 01:21:58.011236 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.011261 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.011546 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.012530 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.017318 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.023152 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.023650 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.023812 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.024209 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.070041 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr"] Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.103738 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.103952 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.104137 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.104308 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2czq\" (UniqueName: \"kubernetes.io/projected/ac32e469-f344-4ee1-8702-bfa8b124ac35-kube-api-access-c2czq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.104448 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.206522 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.206627 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2czq\" (UniqueName: \"kubernetes.io/projected/ac32e469-f344-4ee1-8702-bfa8b124ac35-kube-api-access-c2czq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.206679 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.206727 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.206783 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.211977 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.212390 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.212400 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.214791 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.224518 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2czq\" (UniqueName: \"kubernetes.io/projected/ac32e469-f344-4ee1-8702-bfa8b124ac35-kube-api-access-c2czq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.337535 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:21:58 crc kubenswrapper[4912]: I1203 01:21:58.934569 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr"] Dec 03 01:21:59 crc kubenswrapper[4912]: I1203 01:21:59.956724 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" event={"ID":"ac32e469-f344-4ee1-8702-bfa8b124ac35","Type":"ContainerStarted","Data":"6458b6e2361e4246165c48bcd2da9a3b211f8adfdd022161c3d94ba59284f59b"} Dec 03 01:21:59 crc kubenswrapper[4912]: I1203 01:21:59.957184 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" event={"ID":"ac32e469-f344-4ee1-8702-bfa8b124ac35","Type":"ContainerStarted","Data":"93c41539365d874bbe12d3ba581bccb4d962d7d2c113d4c364b8d24bf8aa801c"} Dec 03 01:22:00 crc kubenswrapper[4912]: I1203 01:22:00.015226 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" podStartSLOduration=2.462376203 podStartE2EDuration="3.015200104s" podCreationTimestamp="2025-12-03 01:21:57 +0000 UTC" firstStartedPulling="2025-12-03 01:21:58.950373631 +0000 UTC m=+3504.592394221" lastFinishedPulling="2025-12-03 01:21:59.503197532 +0000 UTC m=+3505.145218122" observedRunningTime="2025-12-03 01:22:00.00178191 +0000 UTC m=+3505.643802480" watchObservedRunningTime="2025-12-03 01:22:00.015200104 +0000 UTC m=+3505.657220704" Dec 03 01:22:23 crc kubenswrapper[4912]: I1203 01:22:23.559371 4912 scope.go:117] "RemoveContainer" containerID="cc3096107788de126604dc147eed2f7c40312aa523a426f83a7f8352aab9926b" Dec 03 01:22:23 crc kubenswrapper[4912]: I1203 01:22:23.611401 4912 scope.go:117] "RemoveContainer" containerID="bea415e1b94c3e8534e5b60c11e684118664e29409237c7273fd3e5efadebad0" Dec 03 01:22:23 crc kubenswrapper[4912]: I1203 01:22:23.693820 4912 scope.go:117] "RemoveContainer" containerID="b1ac1f75e00f07f6f097bf0b24de3ff3d53d3ea17916c7d0a7c7047cdfc2d665" Dec 03 01:22:23 crc kubenswrapper[4912]: I1203 01:22:23.754643 4912 scope.go:117] "RemoveContainer" containerID="d89f010efa30b12367cce5e47a57d3a2aa8b796588c87fc9f5bc4d3ab46c953a" Dec 03 01:22:23 crc kubenswrapper[4912]: I1203 01:22:23.802310 4912 scope.go:117] "RemoveContainer" containerID="04b81e41c9836381a227ffbc8251c0b5ca9855f82315cb9631ec842a805200d5" Dec 03 01:22:23 crc kubenswrapper[4912]: I1203 01:22:23.862342 4912 scope.go:117] "RemoveContainer" containerID="5c9d9eff1acf7d358d730edc5e20382faa3aca1799e3ace928e84fadea14bc80" Dec 03 01:22:23 crc kubenswrapper[4912]: I1203 01:22:23.932281 4912 scope.go:117] "RemoveContainer" containerID="ab55aab0ad54779688c3948006ebfb6e69d5a5689c75f8b47a6eb20424b7df3a" Dec 03 01:22:23 crc kubenswrapper[4912]: I1203 01:22:23.971085 4912 scope.go:117] "RemoveContainer" containerID="76942f72ae2455ee656d3d27d0d69e7868590808d99ca5e616938969d6beb5c3" Dec 03 01:22:24 crc kubenswrapper[4912]: I1203 01:22:24.047669 4912 scope.go:117] "RemoveContainer" containerID="9cf1018ed6c7597b191dc7ee1c7b36e89de4314ea203e76c7f54838a2f8209f6" Dec 03 01:22:24 crc kubenswrapper[4912]: I1203 01:22:24.100361 4912 scope.go:117] "RemoveContainer" containerID="cfc8a87e1b7079bf7027d26e0a73fee81bb31d89e7b23ffbdf5ead97447e2c4e" Dec 03 01:22:24 crc kubenswrapper[4912]: I1203 01:22:24.181906 4912 scope.go:117] "RemoveContainer" containerID="7c124d8dc3ea3a8de90f863adbac4f38fe8cf8145b9ed6605f57f95dece85c30" Dec 03 01:22:24 crc kubenswrapper[4912]: I1203 01:22:24.277729 4912 scope.go:117] "RemoveContainer" containerID="303c3fac11f3c5d54968f50f704d4a10d63349222450e789008ca84a96eca627" Dec 03 01:22:24 crc kubenswrapper[4912]: I1203 01:22:24.395302 4912 scope.go:117] "RemoveContainer" containerID="e5bbe8ee1180ba8b1c51293564b93745cb0c2d026d0487f9e167c4fc3448c7c5" Dec 03 01:22:24 crc kubenswrapper[4912]: I1203 01:22:24.465243 4912 scope.go:117] "RemoveContainer" containerID="f91418c182e59bd5f6225519590b349d8193d859b3d08c0a4d07864c1de304e5" Dec 03 01:22:24 crc kubenswrapper[4912]: I1203 01:22:24.513036 4912 scope.go:117] "RemoveContainer" containerID="7aab6c5a1fa5d9a4cd1835efbff54d2dc216b26997bbe60f483abc7da0765283" Dec 03 01:22:24 crc kubenswrapper[4912]: I1203 01:22:24.585368 4912 scope.go:117] "RemoveContainer" containerID="8906b82ec139ea8f3d6e0d906c1bf93b2cac357db6496ee4e9b181c9caea469b" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.320057 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vtxkv"] Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.326174 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.332070 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtxkv"] Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.426842 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnnnl\" (UniqueName: \"kubernetes.io/projected/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-kube-api-access-pnnnl\") pod \"redhat-marketplace-vtxkv\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.426929 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-catalog-content\") pod \"redhat-marketplace-vtxkv\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.427089 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-utilities\") pod \"redhat-marketplace-vtxkv\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.528100 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-utilities\") pod \"redhat-marketplace-vtxkv\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.528213 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnnnl\" (UniqueName: \"kubernetes.io/projected/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-kube-api-access-pnnnl\") pod \"redhat-marketplace-vtxkv\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.528260 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-catalog-content\") pod \"redhat-marketplace-vtxkv\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.528664 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-utilities\") pod \"redhat-marketplace-vtxkv\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.529036 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-catalog-content\") pod \"redhat-marketplace-vtxkv\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.548020 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnnnl\" (UniqueName: \"kubernetes.io/projected/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-kube-api-access-pnnnl\") pod \"redhat-marketplace-vtxkv\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:26 crc kubenswrapper[4912]: I1203 01:22:26.654153 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:27 crc kubenswrapper[4912]: I1203 01:22:27.161745 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtxkv"] Dec 03 01:22:27 crc kubenswrapper[4912]: I1203 01:22:27.378620 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtxkv" event={"ID":"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a","Type":"ContainerStarted","Data":"fbc093ac19121a56080164604ee87ccf895cfab2fd6513e04d198bfeeebc7200"} Dec 03 01:22:28 crc kubenswrapper[4912]: I1203 01:22:28.393503 4912 generic.go:334] "Generic (PLEG): container finished" podID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerID="0b97aee6ad17866cb118d12e6a09dcdcf4bdce99a2635df8f65a10d7e64c82d0" exitCode=0 Dec 03 01:22:28 crc kubenswrapper[4912]: I1203 01:22:28.393600 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtxkv" event={"ID":"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a","Type":"ContainerDied","Data":"0b97aee6ad17866cb118d12e6a09dcdcf4bdce99a2635df8f65a10d7e64c82d0"} Dec 03 01:22:30 crc kubenswrapper[4912]: I1203 01:22:30.417719 4912 generic.go:334] "Generic (PLEG): container finished" podID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerID="6ad7bafc8f3e0b532b49f7fb40f87802b84d36c6fa034b5d65f0e47d93053c7b" exitCode=0 Dec 03 01:22:30 crc kubenswrapper[4912]: I1203 01:22:30.417776 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtxkv" event={"ID":"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a","Type":"ContainerDied","Data":"6ad7bafc8f3e0b532b49f7fb40f87802b84d36c6fa034b5d65f0e47d93053c7b"} Dec 03 01:22:31 crc kubenswrapper[4912]: I1203 01:22:31.431953 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtxkv" event={"ID":"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a","Type":"ContainerStarted","Data":"f2d218f7f1c30881d3142245f73d3151986af61588fe17106c33d3ddfc08e9b9"} Dec 03 01:22:36 crc kubenswrapper[4912]: I1203 01:22:36.656235 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:36 crc kubenswrapper[4912]: I1203 01:22:36.656805 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:36 crc kubenswrapper[4912]: I1203 01:22:36.753165 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:36 crc kubenswrapper[4912]: I1203 01:22:36.780085 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vtxkv" podStartSLOduration=8.281216235 podStartE2EDuration="10.780069531s" podCreationTimestamp="2025-12-03 01:22:26 +0000 UTC" firstStartedPulling="2025-12-03 01:22:28.39724476 +0000 UTC m=+3534.039265360" lastFinishedPulling="2025-12-03 01:22:30.896098066 +0000 UTC m=+3536.538118656" observedRunningTime="2025-12-03 01:22:31.472173052 +0000 UTC m=+3537.114193612" watchObservedRunningTime="2025-12-03 01:22:36.780069531 +0000 UTC m=+3542.422090091" Dec 03 01:22:37 crc kubenswrapper[4912]: I1203 01:22:37.568711 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:37 crc kubenswrapper[4912]: I1203 01:22:37.630872 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtxkv"] Dec 03 01:22:39 crc kubenswrapper[4912]: I1203 01:22:39.533314 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vtxkv" podUID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerName="registry-server" containerID="cri-o://f2d218f7f1c30881d3142245f73d3151986af61588fe17106c33d3ddfc08e9b9" gracePeriod=2 Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.560191 4912 generic.go:334] "Generic (PLEG): container finished" podID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerID="f2d218f7f1c30881d3142245f73d3151986af61588fe17106c33d3ddfc08e9b9" exitCode=0 Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.560251 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtxkv" event={"ID":"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a","Type":"ContainerDied","Data":"f2d218f7f1c30881d3142245f73d3151986af61588fe17106c33d3ddfc08e9b9"} Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.764084 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.894998 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-utilities\") pod \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.895244 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnnnl\" (UniqueName: \"kubernetes.io/projected/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-kube-api-access-pnnnl\") pod \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.895276 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-catalog-content\") pod \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\" (UID: \"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a\") " Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.898794 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-utilities" (OuterVolumeSpecName: "utilities") pod "5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" (UID: "5ea86cff-5fd5-440f-acf5-dc7e26d95b3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.905652 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-kube-api-access-pnnnl" (OuterVolumeSpecName: "kube-api-access-pnnnl") pod "5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" (UID: "5ea86cff-5fd5-440f-acf5-dc7e26d95b3a"). InnerVolumeSpecName "kube-api-access-pnnnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.921393 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" (UID: "5ea86cff-5fd5-440f-acf5-dc7e26d95b3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.997325 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.997365 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnnnl\" (UniqueName: \"kubernetes.io/projected/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-kube-api-access-pnnnl\") on node \"crc\" DevicePath \"\"" Dec 03 01:22:40 crc kubenswrapper[4912]: I1203 01:22:40.997380 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:22:41 crc kubenswrapper[4912]: I1203 01:22:41.577317 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vtxkv" event={"ID":"5ea86cff-5fd5-440f-acf5-dc7e26d95b3a","Type":"ContainerDied","Data":"fbc093ac19121a56080164604ee87ccf895cfab2fd6513e04d198bfeeebc7200"} Dec 03 01:22:41 crc kubenswrapper[4912]: I1203 01:22:41.577618 4912 scope.go:117] "RemoveContainer" containerID="f2d218f7f1c30881d3142245f73d3151986af61588fe17106c33d3ddfc08e9b9" Dec 03 01:22:41 crc kubenswrapper[4912]: I1203 01:22:41.577470 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vtxkv" Dec 03 01:22:41 crc kubenswrapper[4912]: I1203 01:22:41.624781 4912 scope.go:117] "RemoveContainer" containerID="6ad7bafc8f3e0b532b49f7fb40f87802b84d36c6fa034b5d65f0e47d93053c7b" Dec 03 01:22:41 crc kubenswrapper[4912]: I1203 01:22:41.634223 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtxkv"] Dec 03 01:22:41 crc kubenswrapper[4912]: I1203 01:22:41.658318 4912 scope.go:117] "RemoveContainer" containerID="0b97aee6ad17866cb118d12e6a09dcdcf4bdce99a2635df8f65a10d7e64c82d0" Dec 03 01:22:41 crc kubenswrapper[4912]: I1203 01:22:41.660739 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vtxkv"] Dec 03 01:22:42 crc kubenswrapper[4912]: I1203 01:22:42.593199 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" path="/var/lib/kubelet/pods/5ea86cff-5fd5-440f-acf5-dc7e26d95b3a/volumes" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.077734 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.078229 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.131021 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c4cs6"] Dec 03 01:23:18 crc kubenswrapper[4912]: E1203 01:23:18.131647 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerName="extract-utilities" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.131663 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerName="extract-utilities" Dec 03 01:23:18 crc kubenswrapper[4912]: E1203 01:23:18.131681 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerName="extract-content" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.131689 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerName="extract-content" Dec 03 01:23:18 crc kubenswrapper[4912]: E1203 01:23:18.131697 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerName="registry-server" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.131705 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerName="registry-server" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.131908 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ea86cff-5fd5-440f-acf5-dc7e26d95b3a" containerName="registry-server" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.133334 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.168955 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4cs6"] Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.193078 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-utilities\") pod \"redhat-operators-c4cs6\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.193212 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spfrf\" (UniqueName: \"kubernetes.io/projected/cc019676-f8df-4a4c-9b72-87d3240c39a1-kube-api-access-spfrf\") pod \"redhat-operators-c4cs6\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.193258 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-catalog-content\") pod \"redhat-operators-c4cs6\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.295766 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spfrf\" (UniqueName: \"kubernetes.io/projected/cc019676-f8df-4a4c-9b72-87d3240c39a1-kube-api-access-spfrf\") pod \"redhat-operators-c4cs6\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.295817 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-catalog-content\") pod \"redhat-operators-c4cs6\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.296001 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-utilities\") pod \"redhat-operators-c4cs6\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.296624 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-utilities\") pod \"redhat-operators-c4cs6\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.297153 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-catalog-content\") pod \"redhat-operators-c4cs6\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.319511 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spfrf\" (UniqueName: \"kubernetes.io/projected/cc019676-f8df-4a4c-9b72-87d3240c39a1-kube-api-access-spfrf\") pod \"redhat-operators-c4cs6\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.463005 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:18 crc kubenswrapper[4912]: I1203 01:23:18.957930 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c4cs6"] Dec 03 01:23:19 crc kubenswrapper[4912]: I1203 01:23:19.033178 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cs6" event={"ID":"cc019676-f8df-4a4c-9b72-87d3240c39a1","Type":"ContainerStarted","Data":"3dd0e6315d862839319fd6684f1e5155e6d6d8329063c279cd6d938006070b30"} Dec 03 01:23:20 crc kubenswrapper[4912]: I1203 01:23:20.046654 4912 generic.go:334] "Generic (PLEG): container finished" podID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerID="d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff" exitCode=0 Dec 03 01:23:20 crc kubenswrapper[4912]: I1203 01:23:20.046777 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cs6" event={"ID":"cc019676-f8df-4a4c-9b72-87d3240c39a1","Type":"ContainerDied","Data":"d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff"} Dec 03 01:23:22 crc kubenswrapper[4912]: I1203 01:23:22.070481 4912 generic.go:334] "Generic (PLEG): container finished" podID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerID="05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d" exitCode=0 Dec 03 01:23:22 crc kubenswrapper[4912]: I1203 01:23:22.070951 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cs6" event={"ID":"cc019676-f8df-4a4c-9b72-87d3240c39a1","Type":"ContainerDied","Data":"05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d"} Dec 03 01:23:24 crc kubenswrapper[4912]: I1203 01:23:24.100836 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cs6" event={"ID":"cc019676-f8df-4a4c-9b72-87d3240c39a1","Type":"ContainerStarted","Data":"e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5"} Dec 03 01:23:24 crc kubenswrapper[4912]: I1203 01:23:24.147197 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c4cs6" podStartSLOduration=3.553002713 podStartE2EDuration="6.147176839s" podCreationTimestamp="2025-12-03 01:23:18 +0000 UTC" firstStartedPulling="2025-12-03 01:23:20.050844732 +0000 UTC m=+3585.692865302" lastFinishedPulling="2025-12-03 01:23:22.645018868 +0000 UTC m=+3588.287039428" observedRunningTime="2025-12-03 01:23:24.130255782 +0000 UTC m=+3589.772276352" watchObservedRunningTime="2025-12-03 01:23:24.147176839 +0000 UTC m=+3589.789197409" Dec 03 01:23:28 crc kubenswrapper[4912]: I1203 01:23:28.463323 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:28 crc kubenswrapper[4912]: I1203 01:23:28.463982 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:29 crc kubenswrapper[4912]: I1203 01:23:29.519682 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c4cs6" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerName="registry-server" probeResult="failure" output=< Dec 03 01:23:29 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 01:23:29 crc kubenswrapper[4912]: > Dec 03 01:23:38 crc kubenswrapper[4912]: I1203 01:23:38.550058 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:38 crc kubenswrapper[4912]: I1203 01:23:38.629978 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:38 crc kubenswrapper[4912]: I1203 01:23:38.811355 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c4cs6"] Dec 03 01:23:40 crc kubenswrapper[4912]: I1203 01:23:40.307974 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c4cs6" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerName="registry-server" containerID="cri-o://e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5" gracePeriod=2 Dec 03 01:23:40 crc kubenswrapper[4912]: I1203 01:23:40.854850 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:40 crc kubenswrapper[4912]: I1203 01:23:40.967259 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-catalog-content\") pod \"cc019676-f8df-4a4c-9b72-87d3240c39a1\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " Dec 03 01:23:40 crc kubenswrapper[4912]: I1203 01:23:40.967654 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-utilities\") pod \"cc019676-f8df-4a4c-9b72-87d3240c39a1\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " Dec 03 01:23:40 crc kubenswrapper[4912]: I1203 01:23:40.967805 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spfrf\" (UniqueName: \"kubernetes.io/projected/cc019676-f8df-4a4c-9b72-87d3240c39a1-kube-api-access-spfrf\") pod \"cc019676-f8df-4a4c-9b72-87d3240c39a1\" (UID: \"cc019676-f8df-4a4c-9b72-87d3240c39a1\") " Dec 03 01:23:40 crc kubenswrapper[4912]: I1203 01:23:40.968893 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-utilities" (OuterVolumeSpecName: "utilities") pod "cc019676-f8df-4a4c-9b72-87d3240c39a1" (UID: "cc019676-f8df-4a4c-9b72-87d3240c39a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:23:40 crc kubenswrapper[4912]: I1203 01:23:40.977985 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc019676-f8df-4a4c-9b72-87d3240c39a1-kube-api-access-spfrf" (OuterVolumeSpecName: "kube-api-access-spfrf") pod "cc019676-f8df-4a4c-9b72-87d3240c39a1" (UID: "cc019676-f8df-4a4c-9b72-87d3240c39a1"). InnerVolumeSpecName "kube-api-access-spfrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.070606 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.070894 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spfrf\" (UniqueName: \"kubernetes.io/projected/cc019676-f8df-4a4c-9b72-87d3240c39a1-kube-api-access-spfrf\") on node \"crc\" DevicePath \"\"" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.095686 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc019676-f8df-4a4c-9b72-87d3240c39a1" (UID: "cc019676-f8df-4a4c-9b72-87d3240c39a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.173354 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc019676-f8df-4a4c-9b72-87d3240c39a1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.321378 4912 generic.go:334] "Generic (PLEG): container finished" podID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerID="e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5" exitCode=0 Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.321485 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c4cs6" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.322520 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cs6" event={"ID":"cc019676-f8df-4a4c-9b72-87d3240c39a1","Type":"ContainerDied","Data":"e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5"} Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.322565 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c4cs6" event={"ID":"cc019676-f8df-4a4c-9b72-87d3240c39a1","Type":"ContainerDied","Data":"3dd0e6315d862839319fd6684f1e5155e6d6d8329063c279cd6d938006070b30"} Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.322597 4912 scope.go:117] "RemoveContainer" containerID="e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.363210 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c4cs6"] Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.364881 4912 scope.go:117] "RemoveContainer" containerID="05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.372601 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c4cs6"] Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.406877 4912 scope.go:117] "RemoveContainer" containerID="d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.475800 4912 scope.go:117] "RemoveContainer" containerID="e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5" Dec 03 01:23:41 crc kubenswrapper[4912]: E1203 01:23:41.477234 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5\": container with ID starting with e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5 not found: ID does not exist" containerID="e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.477278 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5"} err="failed to get container status \"e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5\": rpc error: code = NotFound desc = could not find container \"e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5\": container with ID starting with e07f4036e3b1069136f3fd58c6708305b18a6d20d457f41784933fef3f8166b5 not found: ID does not exist" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.477303 4912 scope.go:117] "RemoveContainer" containerID="05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d" Dec 03 01:23:41 crc kubenswrapper[4912]: E1203 01:23:41.479256 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d\": container with ID starting with 05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d not found: ID does not exist" containerID="05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.479314 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d"} err="failed to get container status \"05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d\": rpc error: code = NotFound desc = could not find container \"05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d\": container with ID starting with 05614a26f9932eeb6e9b990151637fd58d345589ba2a979c2a4e3c3adf645b4d not found: ID does not exist" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.479345 4912 scope.go:117] "RemoveContainer" containerID="d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff" Dec 03 01:23:41 crc kubenswrapper[4912]: E1203 01:23:41.480029 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff\": container with ID starting with d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff not found: ID does not exist" containerID="d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff" Dec 03 01:23:41 crc kubenswrapper[4912]: I1203 01:23:41.480097 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff"} err="failed to get container status \"d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff\": rpc error: code = NotFound desc = could not find container \"d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff\": container with ID starting with d156a4e35fd20fcda14a10be7c3448f6b0af03109705c70206ce54c3356359ff not found: ID does not exist" Dec 03 01:23:42 crc kubenswrapper[4912]: I1203 01:23:42.592386 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" path="/var/lib/kubelet/pods/cc019676-f8df-4a4c-9b72-87d3240c39a1/volumes" Dec 03 01:23:48 crc kubenswrapper[4912]: I1203 01:23:48.078248 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:23:48 crc kubenswrapper[4912]: I1203 01:23:48.079101 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:23:57 crc kubenswrapper[4912]: I1203 01:23:57.510793 4912 generic.go:334] "Generic (PLEG): container finished" podID="ac32e469-f344-4ee1-8702-bfa8b124ac35" containerID="6458b6e2361e4246165c48bcd2da9a3b211f8adfdd022161c3d94ba59284f59b" exitCode=0 Dec 03 01:23:57 crc kubenswrapper[4912]: I1203 01:23:57.510910 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" event={"ID":"ac32e469-f344-4ee1-8702-bfa8b124ac35","Type":"ContainerDied","Data":"6458b6e2361e4246165c48bcd2da9a3b211f8adfdd022161c3d94ba59284f59b"} Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.050341 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.230478 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2czq\" (UniqueName: \"kubernetes.io/projected/ac32e469-f344-4ee1-8702-bfa8b124ac35-kube-api-access-c2czq\") pod \"ac32e469-f344-4ee1-8702-bfa8b124ac35\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.230565 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-inventory\") pod \"ac32e469-f344-4ee1-8702-bfa8b124ac35\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.230675 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-bootstrap-combined-ca-bundle\") pod \"ac32e469-f344-4ee1-8702-bfa8b124ac35\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.230713 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ceph\") pod \"ac32e469-f344-4ee1-8702-bfa8b124ac35\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.230772 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ssh-key\") pod \"ac32e469-f344-4ee1-8702-bfa8b124ac35\" (UID: \"ac32e469-f344-4ee1-8702-bfa8b124ac35\") " Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.244274 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ceph" (OuterVolumeSpecName: "ceph") pod "ac32e469-f344-4ee1-8702-bfa8b124ac35" (UID: "ac32e469-f344-4ee1-8702-bfa8b124ac35"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.246562 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac32e469-f344-4ee1-8702-bfa8b124ac35-kube-api-access-c2czq" (OuterVolumeSpecName: "kube-api-access-c2czq") pod "ac32e469-f344-4ee1-8702-bfa8b124ac35" (UID: "ac32e469-f344-4ee1-8702-bfa8b124ac35"). InnerVolumeSpecName "kube-api-access-c2czq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.246583 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "ac32e469-f344-4ee1-8702-bfa8b124ac35" (UID: "ac32e469-f344-4ee1-8702-bfa8b124ac35"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.273918 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ac32e469-f344-4ee1-8702-bfa8b124ac35" (UID: "ac32e469-f344-4ee1-8702-bfa8b124ac35"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.296251 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-inventory" (OuterVolumeSpecName: "inventory") pod "ac32e469-f344-4ee1-8702-bfa8b124ac35" (UID: "ac32e469-f344-4ee1-8702-bfa8b124ac35"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.335014 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2czq\" (UniqueName: \"kubernetes.io/projected/ac32e469-f344-4ee1-8702-bfa8b124ac35-kube-api-access-c2czq\") on node \"crc\" DevicePath \"\"" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.335066 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.335083 4912 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.335098 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.335113 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac32e469-f344-4ee1-8702-bfa8b124ac35-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.534096 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" event={"ID":"ac32e469-f344-4ee1-8702-bfa8b124ac35","Type":"ContainerDied","Data":"93c41539365d874bbe12d3ba581bccb4d962d7d2c113d4c364b8d24bf8aa801c"} Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.534139 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.534149 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93c41539365d874bbe12d3ba581bccb4d962d7d2c113d4c364b8d24bf8aa801c" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.656775 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv"] Dec 03 01:23:59 crc kubenswrapper[4912]: E1203 01:23:59.657319 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerName="extract-content" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.657346 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerName="extract-content" Dec 03 01:23:59 crc kubenswrapper[4912]: E1203 01:23:59.657387 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerName="registry-server" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.657395 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerName="registry-server" Dec 03 01:23:59 crc kubenswrapper[4912]: E1203 01:23:59.657412 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerName="extract-utilities" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.657420 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerName="extract-utilities" Dec 03 01:23:59 crc kubenswrapper[4912]: E1203 01:23:59.657603 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac32e469-f344-4ee1-8702-bfa8b124ac35" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.657623 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac32e469-f344-4ee1-8702-bfa8b124ac35" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.658140 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc019676-f8df-4a4c-9b72-87d3240c39a1" containerName="registry-server" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.658209 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac32e469-f344-4ee1-8702-bfa8b124ac35" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.659075 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.661291 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.662360 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.663747 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.665266 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.666630 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.686783 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv"] Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.744691 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qtdx\" (UniqueName: \"kubernetes.io/projected/50268f51-4c6d-4837-8806-3f741ee7b48c-kube-api-access-8qtdx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.744838 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.745269 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.745436 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.847597 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.848021 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.848334 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qtdx\" (UniqueName: \"kubernetes.io/projected/50268f51-4c6d-4837-8806-3f741ee7b48c-kube-api-access-8qtdx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.848749 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.852564 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.858813 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.859972 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.871532 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qtdx\" (UniqueName: \"kubernetes.io/projected/50268f51-4c6d-4837-8806-3f741ee7b48c-kube-api-access-8qtdx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:23:59 crc kubenswrapper[4912]: I1203 01:23:59.978023 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:24:00 crc kubenswrapper[4912]: I1203 01:24:00.559092 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv"] Dec 03 01:24:00 crc kubenswrapper[4912]: W1203 01:24:00.563833 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50268f51_4c6d_4837_8806_3f741ee7b48c.slice/crio-da524dd85579a9b6b4f6311748692dd98b4b5f4c9debe2340eea66693e1a1157 WatchSource:0}: Error finding container da524dd85579a9b6b4f6311748692dd98b4b5f4c9debe2340eea66693e1a1157: Status 404 returned error can't find the container with id da524dd85579a9b6b4f6311748692dd98b4b5f4c9debe2340eea66693e1a1157 Dec 03 01:24:01 crc kubenswrapper[4912]: I1203 01:24:01.571999 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" event={"ID":"50268f51-4c6d-4837-8806-3f741ee7b48c","Type":"ContainerStarted","Data":"fcf1ae68e0f259fe0b93b18fa7b71792bbb3279b14da9a192a52c73453a28b42"} Dec 03 01:24:01 crc kubenswrapper[4912]: I1203 01:24:01.572374 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" event={"ID":"50268f51-4c6d-4837-8806-3f741ee7b48c","Type":"ContainerStarted","Data":"da524dd85579a9b6b4f6311748692dd98b4b5f4c9debe2340eea66693e1a1157"} Dec 03 01:24:01 crc kubenswrapper[4912]: I1203 01:24:01.596331 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" podStartSLOduration=1.992531544 podStartE2EDuration="2.596313032s" podCreationTimestamp="2025-12-03 01:23:59 +0000 UTC" firstStartedPulling="2025-12-03 01:24:00.565823706 +0000 UTC m=+3626.207844286" lastFinishedPulling="2025-12-03 01:24:01.169605174 +0000 UTC m=+3626.811625774" observedRunningTime="2025-12-03 01:24:01.589049931 +0000 UTC m=+3627.231070511" watchObservedRunningTime="2025-12-03 01:24:01.596313032 +0000 UTC m=+3627.238333592" Dec 03 01:24:18 crc kubenswrapper[4912]: I1203 01:24:18.078321 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:24:18 crc kubenswrapper[4912]: I1203 01:24:18.079126 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:24:18 crc kubenswrapper[4912]: I1203 01:24:18.079176 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:24:18 crc kubenswrapper[4912]: I1203 01:24:18.080147 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:24:18 crc kubenswrapper[4912]: I1203 01:24:18.080205 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" gracePeriod=600 Dec 03 01:24:18 crc kubenswrapper[4912]: E1203 01:24:18.201403 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:24:18 crc kubenswrapper[4912]: I1203 01:24:18.768711 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" exitCode=0 Dec 03 01:24:18 crc kubenswrapper[4912]: I1203 01:24:18.768782 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2"} Dec 03 01:24:18 crc kubenswrapper[4912]: I1203 01:24:18.768823 4912 scope.go:117] "RemoveContainer" containerID="ea5dabf8afee7141982885ac03babc8ac23410c3b3d263c54df4ae0e2cbbdf30" Dec 03 01:24:18 crc kubenswrapper[4912]: I1203 01:24:18.770121 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:24:18 crc kubenswrapper[4912]: E1203 01:24:18.770817 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:24:31 crc kubenswrapper[4912]: I1203 01:24:31.571943 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:24:31 crc kubenswrapper[4912]: E1203 01:24:31.573069 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:24:35 crc kubenswrapper[4912]: I1203 01:24:35.987918 4912 generic.go:334] "Generic (PLEG): container finished" podID="50268f51-4c6d-4837-8806-3f741ee7b48c" containerID="fcf1ae68e0f259fe0b93b18fa7b71792bbb3279b14da9a192a52c73453a28b42" exitCode=0 Dec 03 01:24:35 crc kubenswrapper[4912]: I1203 01:24:35.988024 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" event={"ID":"50268f51-4c6d-4837-8806-3f741ee7b48c","Type":"ContainerDied","Data":"fcf1ae68e0f259fe0b93b18fa7b71792bbb3279b14da9a192a52c73453a28b42"} Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.627336 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.766321 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qtdx\" (UniqueName: \"kubernetes.io/projected/50268f51-4c6d-4837-8806-3f741ee7b48c-kube-api-access-8qtdx\") pod \"50268f51-4c6d-4837-8806-3f741ee7b48c\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.766400 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ssh-key\") pod \"50268f51-4c6d-4837-8806-3f741ee7b48c\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.766642 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ceph\") pod \"50268f51-4c6d-4837-8806-3f741ee7b48c\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.766690 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-inventory\") pod \"50268f51-4c6d-4837-8806-3f741ee7b48c\" (UID: \"50268f51-4c6d-4837-8806-3f741ee7b48c\") " Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.784100 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ceph" (OuterVolumeSpecName: "ceph") pod "50268f51-4c6d-4837-8806-3f741ee7b48c" (UID: "50268f51-4c6d-4837-8806-3f741ee7b48c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.784272 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50268f51-4c6d-4837-8806-3f741ee7b48c-kube-api-access-8qtdx" (OuterVolumeSpecName: "kube-api-access-8qtdx") pod "50268f51-4c6d-4837-8806-3f741ee7b48c" (UID: "50268f51-4c6d-4837-8806-3f741ee7b48c"). InnerVolumeSpecName "kube-api-access-8qtdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.801751 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-inventory" (OuterVolumeSpecName: "inventory") pod "50268f51-4c6d-4837-8806-3f741ee7b48c" (UID: "50268f51-4c6d-4837-8806-3f741ee7b48c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.807641 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "50268f51-4c6d-4837-8806-3f741ee7b48c" (UID: "50268f51-4c6d-4837-8806-3f741ee7b48c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.868736 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qtdx\" (UniqueName: \"kubernetes.io/projected/50268f51-4c6d-4837-8806-3f741ee7b48c-kube-api-access-8qtdx\") on node \"crc\" DevicePath \"\"" Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.868768 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.868778 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:24:37 crc kubenswrapper[4912]: I1203 01:24:37.868786 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50268f51-4c6d-4837-8806-3f741ee7b48c-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.014212 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" event={"ID":"50268f51-4c6d-4837-8806-3f741ee7b48c","Type":"ContainerDied","Data":"da524dd85579a9b6b4f6311748692dd98b4b5f4c9debe2340eea66693e1a1157"} Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.014456 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da524dd85579a9b6b4f6311748692dd98b4b5f4c9debe2340eea66693e1a1157" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.014578 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.135660 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj"] Dec 03 01:24:38 crc kubenswrapper[4912]: E1203 01:24:38.136477 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50268f51-4c6d-4837-8806-3f741ee7b48c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.136510 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="50268f51-4c6d-4837-8806-3f741ee7b48c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.137201 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="50268f51-4c6d-4837-8806-3f741ee7b48c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.138652 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.141210 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.141344 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.141972 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.142065 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.142100 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.163660 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj"] Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.277736 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.278171 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzgtf\" (UniqueName: \"kubernetes.io/projected/078c0c55-2cdc-4164-af87-c60aced0babf-kube-api-access-tzgtf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.278384 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.278481 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.381323 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzgtf\" (UniqueName: \"kubernetes.io/projected/078c0c55-2cdc-4164-af87-c60aced0babf-kube-api-access-tzgtf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.381540 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.381579 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.381665 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.390778 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.393168 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.393975 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.407059 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzgtf\" (UniqueName: \"kubernetes.io/projected/078c0c55-2cdc-4164-af87-c60aced0babf-kube-api-access-tzgtf\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-c75zj\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:38 crc kubenswrapper[4912]: I1203 01:24:38.473826 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:39 crc kubenswrapper[4912]: I1203 01:24:39.101414 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj"] Dec 03 01:24:39 crc kubenswrapper[4912]: W1203 01:24:39.104950 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod078c0c55_2cdc_4164_af87_c60aced0babf.slice/crio-162c93f539162b7deccb3bd2a93f5c7a460611593b08e21a71115f6032895a68 WatchSource:0}: Error finding container 162c93f539162b7deccb3bd2a93f5c7a460611593b08e21a71115f6032895a68: Status 404 returned error can't find the container with id 162c93f539162b7deccb3bd2a93f5c7a460611593b08e21a71115f6032895a68 Dec 03 01:24:40 crc kubenswrapper[4912]: I1203 01:24:40.047981 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" event={"ID":"078c0c55-2cdc-4164-af87-c60aced0babf","Type":"ContainerStarted","Data":"30fab6f09a31322394be276de91c37f028ef775b91369ac762e11addb32dd41f"} Dec 03 01:24:40 crc kubenswrapper[4912]: I1203 01:24:40.048276 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" event={"ID":"078c0c55-2cdc-4164-af87-c60aced0babf","Type":"ContainerStarted","Data":"162c93f539162b7deccb3bd2a93f5c7a460611593b08e21a71115f6032895a68"} Dec 03 01:24:42 crc kubenswrapper[4912]: I1203 01:24:42.573999 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:24:42 crc kubenswrapper[4912]: E1203 01:24:42.575215 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:24:47 crc kubenswrapper[4912]: I1203 01:24:47.128732 4912 generic.go:334] "Generic (PLEG): container finished" podID="078c0c55-2cdc-4164-af87-c60aced0babf" containerID="30fab6f09a31322394be276de91c37f028ef775b91369ac762e11addb32dd41f" exitCode=0 Dec 03 01:24:47 crc kubenswrapper[4912]: I1203 01:24:47.128865 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" event={"ID":"078c0c55-2cdc-4164-af87-c60aced0babf","Type":"ContainerDied","Data":"30fab6f09a31322394be276de91c37f028ef775b91369ac762e11addb32dd41f"} Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.682751 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.747143 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzgtf\" (UniqueName: \"kubernetes.io/projected/078c0c55-2cdc-4164-af87-c60aced0babf-kube-api-access-tzgtf\") pod \"078c0c55-2cdc-4164-af87-c60aced0babf\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.747255 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ssh-key\") pod \"078c0c55-2cdc-4164-af87-c60aced0babf\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.747281 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ceph\") pod \"078c0c55-2cdc-4164-af87-c60aced0babf\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.747361 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-inventory\") pod \"078c0c55-2cdc-4164-af87-c60aced0babf\" (UID: \"078c0c55-2cdc-4164-af87-c60aced0babf\") " Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.754248 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/078c0c55-2cdc-4164-af87-c60aced0babf-kube-api-access-tzgtf" (OuterVolumeSpecName: "kube-api-access-tzgtf") pod "078c0c55-2cdc-4164-af87-c60aced0babf" (UID: "078c0c55-2cdc-4164-af87-c60aced0babf"). InnerVolumeSpecName "kube-api-access-tzgtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.757762 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ceph" (OuterVolumeSpecName: "ceph") pod "078c0c55-2cdc-4164-af87-c60aced0babf" (UID: "078c0c55-2cdc-4164-af87-c60aced0babf"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.779674 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "078c0c55-2cdc-4164-af87-c60aced0babf" (UID: "078c0c55-2cdc-4164-af87-c60aced0babf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.792802 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-inventory" (OuterVolumeSpecName: "inventory") pod "078c0c55-2cdc-4164-af87-c60aced0babf" (UID: "078c0c55-2cdc-4164-af87-c60aced0babf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.850133 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.850171 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzgtf\" (UniqueName: \"kubernetes.io/projected/078c0c55-2cdc-4164-af87-c60aced0babf-kube-api-access-tzgtf\") on node \"crc\" DevicePath \"\"" Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.850187 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:24:48 crc kubenswrapper[4912]: I1203 01:24:48.850200 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/078c0c55-2cdc-4164-af87-c60aced0babf-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.156687 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" event={"ID":"078c0c55-2cdc-4164-af87-c60aced0babf","Type":"ContainerDied","Data":"162c93f539162b7deccb3bd2a93f5c7a460611593b08e21a71115f6032895a68"} Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.156931 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="162c93f539162b7deccb3bd2a93f5c7a460611593b08e21a71115f6032895a68" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.156749 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-c75zj" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.252267 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr"] Dec 03 01:24:49 crc kubenswrapper[4912]: E1203 01:24:49.252877 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078c0c55-2cdc-4164-af87-c60aced0babf" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.252901 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="078c0c55-2cdc-4164-af87-c60aced0babf" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.253149 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="078c0c55-2cdc-4164-af87-c60aced0babf" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.254052 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.258545 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.258591 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.258553 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.258837 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.261636 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.265798 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr"] Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.361520 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.361609 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.361660 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.361715 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnwwp\" (UniqueName: \"kubernetes.io/projected/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-kube-api-access-wnwwp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.463206 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.463296 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.463348 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.463408 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnwwp\" (UniqueName: \"kubernetes.io/projected/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-kube-api-access-wnwwp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.468819 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.478126 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.478175 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.496633 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnwwp\" (UniqueName: \"kubernetes.io/projected/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-kube-api-access-wnwwp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jckxr\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:49 crc kubenswrapper[4912]: I1203 01:24:49.576526 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:24:50 crc kubenswrapper[4912]: I1203 01:24:50.326381 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr"] Dec 03 01:24:50 crc kubenswrapper[4912]: W1203 01:24:50.330651 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90fce4e2_3ec4_469d_a50a_8c8df048c2ab.slice/crio-9a688614ad2ac203e3edc4099beb6af125b8b950d274f3ccf22c2f2b476f07cd WatchSource:0}: Error finding container 9a688614ad2ac203e3edc4099beb6af125b8b950d274f3ccf22c2f2b476f07cd: Status 404 returned error can't find the container with id 9a688614ad2ac203e3edc4099beb6af125b8b950d274f3ccf22c2f2b476f07cd Dec 03 01:24:51 crc kubenswrapper[4912]: I1203 01:24:51.185145 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" event={"ID":"90fce4e2-3ec4-469d-a50a-8c8df048c2ab","Type":"ContainerStarted","Data":"2d6c4774484662c57dcd7ffac5c91727b7ea4ca17aa9b896253160541bab5ea6"} Dec 03 01:24:51 crc kubenswrapper[4912]: I1203 01:24:51.185419 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" event={"ID":"90fce4e2-3ec4-469d-a50a-8c8df048c2ab","Type":"ContainerStarted","Data":"9a688614ad2ac203e3edc4099beb6af125b8b950d274f3ccf22c2f2b476f07cd"} Dec 03 01:24:51 crc kubenswrapper[4912]: I1203 01:24:51.211916 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" podStartSLOduration=1.772791183 podStartE2EDuration="2.211894469s" podCreationTimestamp="2025-12-03 01:24:49 +0000 UTC" firstStartedPulling="2025-12-03 01:24:50.332836745 +0000 UTC m=+3675.974857305" lastFinishedPulling="2025-12-03 01:24:50.771940011 +0000 UTC m=+3676.413960591" observedRunningTime="2025-12-03 01:24:51.202024408 +0000 UTC m=+3676.844045028" watchObservedRunningTime="2025-12-03 01:24:51.211894469 +0000 UTC m=+3676.853915039" Dec 03 01:24:54 crc kubenswrapper[4912]: I1203 01:24:54.590368 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:24:54 crc kubenswrapper[4912]: E1203 01:24:54.591491 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:25:08 crc kubenswrapper[4912]: I1203 01:25:08.571903 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:25:08 crc kubenswrapper[4912]: E1203 01:25:08.572721 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:25:22 crc kubenswrapper[4912]: I1203 01:25:22.573844 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:25:22 crc kubenswrapper[4912]: E1203 01:25:22.574839 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:25:33 crc kubenswrapper[4912]: I1203 01:25:33.572956 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:25:33 crc kubenswrapper[4912]: E1203 01:25:33.574220 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:25:43 crc kubenswrapper[4912]: I1203 01:25:43.871761 4912 generic.go:334] "Generic (PLEG): container finished" podID="90fce4e2-3ec4-469d-a50a-8c8df048c2ab" containerID="2d6c4774484662c57dcd7ffac5c91727b7ea4ca17aa9b896253160541bab5ea6" exitCode=0 Dec 03 01:25:43 crc kubenswrapper[4912]: I1203 01:25:43.871905 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" event={"ID":"90fce4e2-3ec4-469d-a50a-8c8df048c2ab","Type":"ContainerDied","Data":"2d6c4774484662c57dcd7ffac5c91727b7ea4ca17aa9b896253160541bab5ea6"} Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.469400 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.572467 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:25:45 crc kubenswrapper[4912]: E1203 01:25:45.572743 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.592863 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-inventory\") pod \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.593085 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ssh-key\") pod \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.593145 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ceph\") pod \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.593196 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnwwp\" (UniqueName: \"kubernetes.io/projected/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-kube-api-access-wnwwp\") pod \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\" (UID: \"90fce4e2-3ec4-469d-a50a-8c8df048c2ab\") " Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.600039 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-kube-api-access-wnwwp" (OuterVolumeSpecName: "kube-api-access-wnwwp") pod "90fce4e2-3ec4-469d-a50a-8c8df048c2ab" (UID: "90fce4e2-3ec4-469d-a50a-8c8df048c2ab"). InnerVolumeSpecName "kube-api-access-wnwwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.599193 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ceph" (OuterVolumeSpecName: "ceph") pod "90fce4e2-3ec4-469d-a50a-8c8df048c2ab" (UID: "90fce4e2-3ec4-469d-a50a-8c8df048c2ab"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.634337 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-inventory" (OuterVolumeSpecName: "inventory") pod "90fce4e2-3ec4-469d-a50a-8c8df048c2ab" (UID: "90fce4e2-3ec4-469d-a50a-8c8df048c2ab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.659535 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "90fce4e2-3ec4-469d-a50a-8c8df048c2ab" (UID: "90fce4e2-3ec4-469d-a50a-8c8df048c2ab"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.696406 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.696466 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.696481 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.696493 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnwwp\" (UniqueName: \"kubernetes.io/projected/90fce4e2-3ec4-469d-a50a-8c8df048c2ab-kube-api-access-wnwwp\") on node \"crc\" DevicePath \"\"" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.901628 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" event={"ID":"90fce4e2-3ec4-469d-a50a-8c8df048c2ab","Type":"ContainerDied","Data":"9a688614ad2ac203e3edc4099beb6af125b8b950d274f3ccf22c2f2b476f07cd"} Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.902055 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a688614ad2ac203e3edc4099beb6af125b8b950d274f3ccf22c2f2b476f07cd" Dec 03 01:25:45 crc kubenswrapper[4912]: I1203 01:25:45.901750 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jckxr" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.009861 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg"] Dec 03 01:25:46 crc kubenswrapper[4912]: E1203 01:25:46.010588 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90fce4e2-3ec4-469d-a50a-8c8df048c2ab" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.010867 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="90fce4e2-3ec4-469d-a50a-8c8df048c2ab" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.011253 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="90fce4e2-3ec4-469d-a50a-8c8df048c2ab" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.012544 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.016242 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.016280 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.016867 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.017108 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.017946 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.043036 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg"] Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.207534 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.207582 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.207646 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.207721 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2v5h\" (UniqueName: \"kubernetes.io/projected/d241940f-a3d4-4c91-93c7-2149a6e25ae2-kube-api-access-h2v5h\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.309177 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.309260 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2v5h\" (UniqueName: \"kubernetes.io/projected/d241940f-a3d4-4c91-93c7-2149a6e25ae2-kube-api-access-h2v5h\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.309379 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.309399 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.315782 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.316069 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.316464 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.333342 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2v5h\" (UniqueName: \"kubernetes.io/projected/d241940f-a3d4-4c91-93c7-2149a6e25ae2-kube-api-access-h2v5h\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.338196 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:46 crc kubenswrapper[4912]: I1203 01:25:46.904968 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg"] Dec 03 01:25:47 crc kubenswrapper[4912]: I1203 01:25:47.926162 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" event={"ID":"d241940f-a3d4-4c91-93c7-2149a6e25ae2","Type":"ContainerStarted","Data":"0c8c74d75e597746a87a4ef1f7732f0e69f4d62aea74fbe0c00aa5c546056fd6"} Dec 03 01:25:47 crc kubenswrapper[4912]: I1203 01:25:47.926545 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" event={"ID":"d241940f-a3d4-4c91-93c7-2149a6e25ae2","Type":"ContainerStarted","Data":"7550d9d544334212072d055b4468b74cd03ab5571928331e8e377a3de8804103"} Dec 03 01:25:47 crc kubenswrapper[4912]: I1203 01:25:47.953176 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" podStartSLOduration=2.417795467 podStartE2EDuration="2.953157508s" podCreationTimestamp="2025-12-03 01:25:45 +0000 UTC" firstStartedPulling="2025-12-03 01:25:46.911378325 +0000 UTC m=+3732.553398915" lastFinishedPulling="2025-12-03 01:25:47.446740346 +0000 UTC m=+3733.088760956" observedRunningTime="2025-12-03 01:25:47.94018817 +0000 UTC m=+3733.582208760" watchObservedRunningTime="2025-12-03 01:25:47.953157508 +0000 UTC m=+3733.595178068" Dec 03 01:25:53 crc kubenswrapper[4912]: I1203 01:25:53.012644 4912 generic.go:334] "Generic (PLEG): container finished" podID="d241940f-a3d4-4c91-93c7-2149a6e25ae2" containerID="0c8c74d75e597746a87a4ef1f7732f0e69f4d62aea74fbe0c00aa5c546056fd6" exitCode=0 Dec 03 01:25:53 crc kubenswrapper[4912]: I1203 01:25:53.012767 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" event={"ID":"d241940f-a3d4-4c91-93c7-2149a6e25ae2","Type":"ContainerDied","Data":"0c8c74d75e597746a87a4ef1f7732f0e69f4d62aea74fbe0c00aa5c546056fd6"} Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.590358 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.755628 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ceph\") pod \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.755722 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2v5h\" (UniqueName: \"kubernetes.io/projected/d241940f-a3d4-4c91-93c7-2149a6e25ae2-kube-api-access-h2v5h\") pod \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.755790 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ssh-key\") pod \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.755851 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-inventory\") pod \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\" (UID: \"d241940f-a3d4-4c91-93c7-2149a6e25ae2\") " Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.763735 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d241940f-a3d4-4c91-93c7-2149a6e25ae2-kube-api-access-h2v5h" (OuterVolumeSpecName: "kube-api-access-h2v5h") pod "d241940f-a3d4-4c91-93c7-2149a6e25ae2" (UID: "d241940f-a3d4-4c91-93c7-2149a6e25ae2"). InnerVolumeSpecName "kube-api-access-h2v5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.763977 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ceph" (OuterVolumeSpecName: "ceph") pod "d241940f-a3d4-4c91-93c7-2149a6e25ae2" (UID: "d241940f-a3d4-4c91-93c7-2149a6e25ae2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.788684 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d241940f-a3d4-4c91-93c7-2149a6e25ae2" (UID: "d241940f-a3d4-4c91-93c7-2149a6e25ae2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.798547 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-inventory" (OuterVolumeSpecName: "inventory") pod "d241940f-a3d4-4c91-93c7-2149a6e25ae2" (UID: "d241940f-a3d4-4c91-93c7-2149a6e25ae2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.862363 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.862423 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.862467 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d241940f-a3d4-4c91-93c7-2149a6e25ae2-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:25:54 crc kubenswrapper[4912]: I1203 01:25:54.862487 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2v5h\" (UniqueName: \"kubernetes.io/projected/d241940f-a3d4-4c91-93c7-2149a6e25ae2-kube-api-access-h2v5h\") on node \"crc\" DevicePath \"\"" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.039486 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" event={"ID":"d241940f-a3d4-4c91-93c7-2149a6e25ae2","Type":"ContainerDied","Data":"7550d9d544334212072d055b4468b74cd03ab5571928331e8e377a3de8804103"} Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.039526 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.039535 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7550d9d544334212072d055b4468b74cd03ab5571928331e8e377a3de8804103" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.130125 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p"] Dec 03 01:25:55 crc kubenswrapper[4912]: E1203 01:25:55.130806 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d241940f-a3d4-4c91-93c7-2149a6e25ae2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.130832 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d241940f-a3d4-4c91-93c7-2149a6e25ae2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.131117 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d241940f-a3d4-4c91-93c7-2149a6e25ae2" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.132279 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.135269 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.136093 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.136299 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.136383 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.136592 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.150187 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p"] Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.269781 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.269910 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.269957 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.270907 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksk7z\" (UniqueName: \"kubernetes.io/projected/12198d9f-b54f-428d-9a81-7d6d2d44f695-kube-api-access-ksk7z\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.374046 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.374133 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.374204 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksk7z\" (UniqueName: \"kubernetes.io/projected/12198d9f-b54f-428d-9a81-7d6d2d44f695-kube-api-access-ksk7z\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.374340 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.379101 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.379418 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.380205 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.392172 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksk7z\" (UniqueName: \"kubernetes.io/projected/12198d9f-b54f-428d-9a81-7d6d2d44f695-kube-api-access-ksk7z\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-78l9p\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:55 crc kubenswrapper[4912]: I1203 01:25:55.461362 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:25:56 crc kubenswrapper[4912]: W1203 01:25:56.047898 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12198d9f_b54f_428d_9a81_7d6d2d44f695.slice/crio-55e7c01bf3aa258d9e50e09e5261757eaa616378c05700a0e3e683a981cb5353 WatchSource:0}: Error finding container 55e7c01bf3aa258d9e50e09e5261757eaa616378c05700a0e3e683a981cb5353: Status 404 returned error can't find the container with id 55e7c01bf3aa258d9e50e09e5261757eaa616378c05700a0e3e683a981cb5353 Dec 03 01:25:56 crc kubenswrapper[4912]: I1203 01:25:56.049636 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p"] Dec 03 01:25:56 crc kubenswrapper[4912]: I1203 01:25:56.571689 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:25:56 crc kubenswrapper[4912]: E1203 01:25:56.572078 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:25:57 crc kubenswrapper[4912]: I1203 01:25:57.072416 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" event={"ID":"12198d9f-b54f-428d-9a81-7d6d2d44f695","Type":"ContainerStarted","Data":"b7508517ac4ac189ae2f57b30a8d701b0c21e83c83fb0c1f0ac4c4bc3c58fb22"} Dec 03 01:25:57 crc kubenswrapper[4912]: I1203 01:25:57.072781 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" event={"ID":"12198d9f-b54f-428d-9a81-7d6d2d44f695","Type":"ContainerStarted","Data":"55e7c01bf3aa258d9e50e09e5261757eaa616378c05700a0e3e683a981cb5353"} Dec 03 01:25:57 crc kubenswrapper[4912]: I1203 01:25:57.120463 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" podStartSLOduration=1.699380841 podStartE2EDuration="2.12041705s" podCreationTimestamp="2025-12-03 01:25:55 +0000 UTC" firstStartedPulling="2025-12-03 01:25:56.051002195 +0000 UTC m=+3741.693022775" lastFinishedPulling="2025-12-03 01:25:56.472038384 +0000 UTC m=+3742.114058984" observedRunningTime="2025-12-03 01:25:57.090715803 +0000 UTC m=+3742.732736363" watchObservedRunningTime="2025-12-03 01:25:57.12041705 +0000 UTC m=+3742.762437630" Dec 03 01:26:08 crc kubenswrapper[4912]: I1203 01:26:08.572100 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:26:08 crc kubenswrapper[4912]: E1203 01:26:08.573248 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:26:23 crc kubenswrapper[4912]: I1203 01:26:23.572327 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:26:23 crc kubenswrapper[4912]: E1203 01:26:23.573673 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:26:34 crc kubenswrapper[4912]: I1203 01:26:34.591358 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:26:34 crc kubenswrapper[4912]: E1203 01:26:34.593937 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:26:43 crc kubenswrapper[4912]: I1203 01:26:43.795422 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xzzmj"] Dec 03 01:26:43 crc kubenswrapper[4912]: I1203 01:26:43.802155 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:43 crc kubenswrapper[4912]: I1203 01:26:43.811327 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzzmj"] Dec 03 01:26:43 crc kubenswrapper[4912]: I1203 01:26:43.951667 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-catalog-content\") pod \"certified-operators-xzzmj\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:43 crc kubenswrapper[4912]: I1203 01:26:43.952040 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmthm\" (UniqueName: \"kubernetes.io/projected/428a3c49-c15e-4221-bc51-d518ce2ef35f-kube-api-access-hmthm\") pod \"certified-operators-xzzmj\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:43 crc kubenswrapper[4912]: I1203 01:26:43.952747 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-utilities\") pod \"certified-operators-xzzmj\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:44 crc kubenswrapper[4912]: I1203 01:26:44.055334 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-utilities\") pod \"certified-operators-xzzmj\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:44 crc kubenswrapper[4912]: I1203 01:26:44.055796 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-catalog-content\") pod \"certified-operators-xzzmj\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:44 crc kubenswrapper[4912]: I1203 01:26:44.055855 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmthm\" (UniqueName: \"kubernetes.io/projected/428a3c49-c15e-4221-bc51-d518ce2ef35f-kube-api-access-hmthm\") pod \"certified-operators-xzzmj\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:44 crc kubenswrapper[4912]: I1203 01:26:44.056706 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-catalog-content\") pod \"certified-operators-xzzmj\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:44 crc kubenswrapper[4912]: I1203 01:26:44.058128 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-utilities\") pod \"certified-operators-xzzmj\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:44 crc kubenswrapper[4912]: I1203 01:26:44.087737 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmthm\" (UniqueName: \"kubernetes.io/projected/428a3c49-c15e-4221-bc51-d518ce2ef35f-kube-api-access-hmthm\") pod \"certified-operators-xzzmj\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:44 crc kubenswrapper[4912]: I1203 01:26:44.136804 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:44 crc kubenswrapper[4912]: I1203 01:26:44.604332 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzzmj"] Dec 03 01:26:44 crc kubenswrapper[4912]: I1203 01:26:44.678810 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzzmj" event={"ID":"428a3c49-c15e-4221-bc51-d518ce2ef35f","Type":"ContainerStarted","Data":"f5bb6b0eade7516a1e46669af51e5b5cb3a11e4b54d648d07d4c713156fab91d"} Dec 03 01:26:45 crc kubenswrapper[4912]: I1203 01:26:45.694220 4912 generic.go:334] "Generic (PLEG): container finished" podID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerID="5838146184a31e7553db7f0db9004633ce0fc4b73923a1829c23ea8b725bf305" exitCode=0 Dec 03 01:26:45 crc kubenswrapper[4912]: I1203 01:26:45.694301 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzzmj" event={"ID":"428a3c49-c15e-4221-bc51-d518ce2ef35f","Type":"ContainerDied","Data":"5838146184a31e7553db7f0db9004633ce0fc4b73923a1829c23ea8b725bf305"} Dec 03 01:26:45 crc kubenswrapper[4912]: I1203 01:26:45.698694 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:26:46 crc kubenswrapper[4912]: I1203 01:26:46.712051 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzzmj" event={"ID":"428a3c49-c15e-4221-bc51-d518ce2ef35f","Type":"ContainerStarted","Data":"b8ea703087999ce7fba46385652f0a5e751ce30f21a6cf4ff921fbeb349c18bf"} Dec 03 01:26:47 crc kubenswrapper[4912]: I1203 01:26:47.726770 4912 generic.go:334] "Generic (PLEG): container finished" podID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerID="b8ea703087999ce7fba46385652f0a5e751ce30f21a6cf4ff921fbeb349c18bf" exitCode=0 Dec 03 01:26:47 crc kubenswrapper[4912]: I1203 01:26:47.727215 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzzmj" event={"ID":"428a3c49-c15e-4221-bc51-d518ce2ef35f","Type":"ContainerDied","Data":"b8ea703087999ce7fba46385652f0a5e751ce30f21a6cf4ff921fbeb349c18bf"} Dec 03 01:26:48 crc kubenswrapper[4912]: I1203 01:26:48.572757 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:26:48 crc kubenswrapper[4912]: E1203 01:26:48.573891 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:26:48 crc kubenswrapper[4912]: I1203 01:26:48.742733 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzzmj" event={"ID":"428a3c49-c15e-4221-bc51-d518ce2ef35f","Type":"ContainerStarted","Data":"27098b1be5e072fb28fc23e9974687aefb3c124888a993c0394f47228d346e75"} Dec 03 01:26:48 crc kubenswrapper[4912]: I1203 01:26:48.769054 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xzzmj" podStartSLOduration=3.246554692 podStartE2EDuration="5.769026847s" podCreationTimestamp="2025-12-03 01:26:43 +0000 UTC" firstStartedPulling="2025-12-03 01:26:45.698013867 +0000 UTC m=+3791.340034467" lastFinishedPulling="2025-12-03 01:26:48.220486022 +0000 UTC m=+3793.862506622" observedRunningTime="2025-12-03 01:26:48.767603688 +0000 UTC m=+3794.409624288" watchObservedRunningTime="2025-12-03 01:26:48.769026847 +0000 UTC m=+3794.411047437" Dec 03 01:26:54 crc kubenswrapper[4912]: I1203 01:26:54.137052 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:54 crc kubenswrapper[4912]: I1203 01:26:54.137749 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:54 crc kubenswrapper[4912]: I1203 01:26:54.232544 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:54 crc kubenswrapper[4912]: I1203 01:26:54.907426 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:54 crc kubenswrapper[4912]: I1203 01:26:54.994101 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzzmj"] Dec 03 01:26:56 crc kubenswrapper[4912]: I1203 01:26:56.850193 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xzzmj" podUID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerName="registry-server" containerID="cri-o://27098b1be5e072fb28fc23e9974687aefb3c124888a993c0394f47228d346e75" gracePeriod=2 Dec 03 01:26:57 crc kubenswrapper[4912]: I1203 01:26:57.859459 4912 generic.go:334] "Generic (PLEG): container finished" podID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerID="27098b1be5e072fb28fc23e9974687aefb3c124888a993c0394f47228d346e75" exitCode=0 Dec 03 01:26:57 crc kubenswrapper[4912]: I1203 01:26:57.859489 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzzmj" event={"ID":"428a3c49-c15e-4221-bc51-d518ce2ef35f","Type":"ContainerDied","Data":"27098b1be5e072fb28fc23e9974687aefb3c124888a993c0394f47228d346e75"} Dec 03 01:26:57 crc kubenswrapper[4912]: I1203 01:26:57.859717 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzzmj" event={"ID":"428a3c49-c15e-4221-bc51-d518ce2ef35f","Type":"ContainerDied","Data":"f5bb6b0eade7516a1e46669af51e5b5cb3a11e4b54d648d07d4c713156fab91d"} Dec 03 01:26:57 crc kubenswrapper[4912]: I1203 01:26:57.859734 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5bb6b0eade7516a1e46669af51e5b5cb3a11e4b54d648d07d4c713156fab91d" Dec 03 01:26:57 crc kubenswrapper[4912]: I1203 01:26:57.943947 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.039234 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-utilities\") pod \"428a3c49-c15e-4221-bc51-d518ce2ef35f\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.039360 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmthm\" (UniqueName: \"kubernetes.io/projected/428a3c49-c15e-4221-bc51-d518ce2ef35f-kube-api-access-hmthm\") pod \"428a3c49-c15e-4221-bc51-d518ce2ef35f\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.039416 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-catalog-content\") pod \"428a3c49-c15e-4221-bc51-d518ce2ef35f\" (UID: \"428a3c49-c15e-4221-bc51-d518ce2ef35f\") " Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.040301 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-utilities" (OuterVolumeSpecName: "utilities") pod "428a3c49-c15e-4221-bc51-d518ce2ef35f" (UID: "428a3c49-c15e-4221-bc51-d518ce2ef35f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.049632 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/428a3c49-c15e-4221-bc51-d518ce2ef35f-kube-api-access-hmthm" (OuterVolumeSpecName: "kube-api-access-hmthm") pod "428a3c49-c15e-4221-bc51-d518ce2ef35f" (UID: "428a3c49-c15e-4221-bc51-d518ce2ef35f"). InnerVolumeSpecName "kube-api-access-hmthm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.100995 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "428a3c49-c15e-4221-bc51-d518ce2ef35f" (UID: "428a3c49-c15e-4221-bc51-d518ce2ef35f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.142484 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmthm\" (UniqueName: \"kubernetes.io/projected/428a3c49-c15e-4221-bc51-d518ce2ef35f-kube-api-access-hmthm\") on node \"crc\" DevicePath \"\"" Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.142519 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.142529 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/428a3c49-c15e-4221-bc51-d518ce2ef35f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.882936 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzzmj" Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.923023 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzzmj"] Dec 03 01:26:58 crc kubenswrapper[4912]: I1203 01:26:58.940824 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xzzmj"] Dec 03 01:27:00 crc kubenswrapper[4912]: I1203 01:27:00.587890 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="428a3c49-c15e-4221-bc51-d518ce2ef35f" path="/var/lib/kubelet/pods/428a3c49-c15e-4221-bc51-d518ce2ef35f/volumes" Dec 03 01:27:01 crc kubenswrapper[4912]: I1203 01:27:01.572303 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:27:01 crc kubenswrapper[4912]: E1203 01:27:01.572817 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:27:02 crc kubenswrapper[4912]: I1203 01:27:02.947199 4912 generic.go:334] "Generic (PLEG): container finished" podID="12198d9f-b54f-428d-9a81-7d6d2d44f695" containerID="b7508517ac4ac189ae2f57b30a8d701b0c21e83c83fb0c1f0ac4c4bc3c58fb22" exitCode=0 Dec 03 01:27:02 crc kubenswrapper[4912]: I1203 01:27:02.947380 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" event={"ID":"12198d9f-b54f-428d-9a81-7d6d2d44f695","Type":"ContainerDied","Data":"b7508517ac4ac189ae2f57b30a8d701b0c21e83c83fb0c1f0ac4c4bc3c58fb22"} Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.483103 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.592513 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksk7z\" (UniqueName: \"kubernetes.io/projected/12198d9f-b54f-428d-9a81-7d6d2d44f695-kube-api-access-ksk7z\") pod \"12198d9f-b54f-428d-9a81-7d6d2d44f695\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.592628 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ssh-key\") pod \"12198d9f-b54f-428d-9a81-7d6d2d44f695\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.592740 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-inventory\") pod \"12198d9f-b54f-428d-9a81-7d6d2d44f695\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.592761 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ceph\") pod \"12198d9f-b54f-428d-9a81-7d6d2d44f695\" (UID: \"12198d9f-b54f-428d-9a81-7d6d2d44f695\") " Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.598385 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12198d9f-b54f-428d-9a81-7d6d2d44f695-kube-api-access-ksk7z" (OuterVolumeSpecName: "kube-api-access-ksk7z") pod "12198d9f-b54f-428d-9a81-7d6d2d44f695" (UID: "12198d9f-b54f-428d-9a81-7d6d2d44f695"). InnerVolumeSpecName "kube-api-access-ksk7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.598822 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ceph" (OuterVolumeSpecName: "ceph") pod "12198d9f-b54f-428d-9a81-7d6d2d44f695" (UID: "12198d9f-b54f-428d-9a81-7d6d2d44f695"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.627346 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "12198d9f-b54f-428d-9a81-7d6d2d44f695" (UID: "12198d9f-b54f-428d-9a81-7d6d2d44f695"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.634718 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-inventory" (OuterVolumeSpecName: "inventory") pod "12198d9f-b54f-428d-9a81-7d6d2d44f695" (UID: "12198d9f-b54f-428d-9a81-7d6d2d44f695"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.695101 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.695363 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.695472 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/12198d9f-b54f-428d-9a81-7d6d2d44f695-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.695561 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksk7z\" (UniqueName: \"kubernetes.io/projected/12198d9f-b54f-428d-9a81-7d6d2d44f695-kube-api-access-ksk7z\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.975223 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" event={"ID":"12198d9f-b54f-428d-9a81-7d6d2d44f695","Type":"ContainerDied","Data":"55e7c01bf3aa258d9e50e09e5261757eaa616378c05700a0e3e683a981cb5353"} Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.975564 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55e7c01bf3aa258d9e50e09e5261757eaa616378c05700a0e3e683a981cb5353" Dec 03 01:27:04 crc kubenswrapper[4912]: I1203 01:27:04.975312 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-78l9p" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.113976 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-nfq59"] Dec 03 01:27:05 crc kubenswrapper[4912]: E1203 01:27:05.114465 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerName="extract-utilities" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.114483 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerName="extract-utilities" Dec 03 01:27:05 crc kubenswrapper[4912]: E1203 01:27:05.114498 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerName="extract-content" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.114506 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerName="extract-content" Dec 03 01:27:05 crc kubenswrapper[4912]: E1203 01:27:05.114528 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerName="registry-server" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.114536 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerName="registry-server" Dec 03 01:27:05 crc kubenswrapper[4912]: E1203 01:27:05.114571 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12198d9f-b54f-428d-9a81-7d6d2d44f695" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.114580 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="12198d9f-b54f-428d-9a81-7d6d2d44f695" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.114770 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="12198d9f-b54f-428d-9a81-7d6d2d44f695" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.114789 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="428a3c49-c15e-4221-bc51-d518ce2ef35f" containerName="registry-server" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.115544 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.119942 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.120374 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.120678 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.120710 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.120825 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.124609 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-nfq59"] Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.306976 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc7tw\" (UniqueName: \"kubernetes.io/projected/58be2429-a01c-42d9-b773-da6cae52f9b3-kube-api-access-kc7tw\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.307136 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.307169 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.307218 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ceph\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.409296 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.409388 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.409487 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ceph\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.409666 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc7tw\" (UniqueName: \"kubernetes.io/projected/58be2429-a01c-42d9-b773-da6cae52f9b3-kube-api-access-kc7tw\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.413876 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.414246 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ceph\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.415130 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.436100 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc7tw\" (UniqueName: \"kubernetes.io/projected/58be2429-a01c-42d9-b773-da6cae52f9b3-kube-api-access-kc7tw\") pod \"ssh-known-hosts-edpm-deployment-nfq59\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:05 crc kubenswrapper[4912]: I1203 01:27:05.444059 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:06 crc kubenswrapper[4912]: I1203 01:27:06.024454 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-nfq59"] Dec 03 01:27:06 crc kubenswrapper[4912]: I1203 01:27:06.995585 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" event={"ID":"58be2429-a01c-42d9-b773-da6cae52f9b3","Type":"ContainerStarted","Data":"b028905f235c2f99c362a9f2d164e5280424f018d4bae2fee00c5e56a4893cb1"} Dec 03 01:27:06 crc kubenswrapper[4912]: I1203 01:27:06.995915 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" event={"ID":"58be2429-a01c-42d9-b773-da6cae52f9b3","Type":"ContainerStarted","Data":"c43ea58a00e7a3cdebbf164d508c2b23062b843c7d822d6f4d6cf96f89277909"} Dec 03 01:27:07 crc kubenswrapper[4912]: I1203 01:27:07.026027 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" podStartSLOduration=1.309944963 podStartE2EDuration="2.026008368s" podCreationTimestamp="2025-12-03 01:27:05 +0000 UTC" firstStartedPulling="2025-12-03 01:27:06.020935671 +0000 UTC m=+3811.662956261" lastFinishedPulling="2025-12-03 01:27:06.736999106 +0000 UTC m=+3812.379019666" observedRunningTime="2025-12-03 01:27:07.012086194 +0000 UTC m=+3812.654106784" watchObservedRunningTime="2025-12-03 01:27:07.026008368 +0000 UTC m=+3812.668028928" Dec 03 01:27:13 crc kubenswrapper[4912]: I1203 01:27:13.572358 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:27:13 crc kubenswrapper[4912]: E1203 01:27:13.573756 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:27:22 crc kubenswrapper[4912]: I1203 01:27:22.180179 4912 generic.go:334] "Generic (PLEG): container finished" podID="58be2429-a01c-42d9-b773-da6cae52f9b3" containerID="b028905f235c2f99c362a9f2d164e5280424f018d4bae2fee00c5e56a4893cb1" exitCode=0 Dec 03 01:27:22 crc kubenswrapper[4912]: I1203 01:27:22.180343 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" event={"ID":"58be2429-a01c-42d9-b773-da6cae52f9b3","Type":"ContainerDied","Data":"b028905f235c2f99c362a9f2d164e5280424f018d4bae2fee00c5e56a4893cb1"} Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.724236 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.865595 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ceph\") pod \"58be2429-a01c-42d9-b773-da6cae52f9b3\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.866381 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc7tw\" (UniqueName: \"kubernetes.io/projected/58be2429-a01c-42d9-b773-da6cae52f9b3-kube-api-access-kc7tw\") pod \"58be2429-a01c-42d9-b773-da6cae52f9b3\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.866417 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-inventory-0\") pod \"58be2429-a01c-42d9-b773-da6cae52f9b3\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.866786 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ssh-key-openstack-edpm-ipam\") pod \"58be2429-a01c-42d9-b773-da6cae52f9b3\" (UID: \"58be2429-a01c-42d9-b773-da6cae52f9b3\") " Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.871494 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58be2429-a01c-42d9-b773-da6cae52f9b3-kube-api-access-kc7tw" (OuterVolumeSpecName: "kube-api-access-kc7tw") pod "58be2429-a01c-42d9-b773-da6cae52f9b3" (UID: "58be2429-a01c-42d9-b773-da6cae52f9b3"). InnerVolumeSpecName "kube-api-access-kc7tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.872894 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ceph" (OuterVolumeSpecName: "ceph") pod "58be2429-a01c-42d9-b773-da6cae52f9b3" (UID: "58be2429-a01c-42d9-b773-da6cae52f9b3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.927170 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "58be2429-a01c-42d9-b773-da6cae52f9b3" (UID: "58be2429-a01c-42d9-b773-da6cae52f9b3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.929561 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "58be2429-a01c-42d9-b773-da6cae52f9b3" (UID: "58be2429-a01c-42d9-b773-da6cae52f9b3"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.969107 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.969142 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.969163 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc7tw\" (UniqueName: \"kubernetes.io/projected/58be2429-a01c-42d9-b773-da6cae52f9b3-kube-api-access-kc7tw\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:23 crc kubenswrapper[4912]: I1203 01:27:23.969175 4912 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/58be2429-a01c-42d9-b773-da6cae52f9b3-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.199508 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" event={"ID":"58be2429-a01c-42d9-b773-da6cae52f9b3","Type":"ContainerDied","Data":"c43ea58a00e7a3cdebbf164d508c2b23062b843c7d822d6f4d6cf96f89277909"} Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.199551 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c43ea58a00e7a3cdebbf164d508c2b23062b843c7d822d6f4d6cf96f89277909" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.199614 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-nfq59" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.400000 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8"] Dec 03 01:27:24 crc kubenswrapper[4912]: E1203 01:27:24.400458 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58be2429-a01c-42d9-b773-da6cae52f9b3" containerName="ssh-known-hosts-edpm-deployment" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.400472 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="58be2429-a01c-42d9-b773-da6cae52f9b3" containerName="ssh-known-hosts-edpm-deployment" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.400677 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="58be2429-a01c-42d9-b773-da6cae52f9b3" containerName="ssh-known-hosts-edpm-deployment" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.401397 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.407934 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.408082 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.408170 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.408309 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.408336 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.429488 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8"] Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.495952 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.496092 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.496115 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xw7f\" (UniqueName: \"kubernetes.io/projected/d31b18ef-e64f-47ac-9aad-401795f9e3f0-kube-api-access-5xw7f\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.496196 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.598084 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.598208 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.598320 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.598345 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xw7f\" (UniqueName: \"kubernetes.io/projected/d31b18ef-e64f-47ac-9aad-401795f9e3f0-kube-api-access-5xw7f\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.602920 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.605517 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.610859 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.618521 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xw7f\" (UniqueName: \"kubernetes.io/projected/d31b18ef-e64f-47ac-9aad-401795f9e3f0-kube-api-access-5xw7f\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6kzw8\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:24 crc kubenswrapper[4912]: I1203 01:27:24.733193 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:25 crc kubenswrapper[4912]: I1203 01:27:25.384698 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8"] Dec 03 01:27:25 crc kubenswrapper[4912]: W1203 01:27:25.391633 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd31b18ef_e64f_47ac_9aad_401795f9e3f0.slice/crio-676d53c0344e2c7f498f8f489a715047b3a27a1a87bf2cf97f724dc2dfdfed5d WatchSource:0}: Error finding container 676d53c0344e2c7f498f8f489a715047b3a27a1a87bf2cf97f724dc2dfdfed5d: Status 404 returned error can't find the container with id 676d53c0344e2c7f498f8f489a715047b3a27a1a87bf2cf97f724dc2dfdfed5d Dec 03 01:27:25 crc kubenswrapper[4912]: I1203 01:27:25.572349 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:27:25 crc kubenswrapper[4912]: E1203 01:27:25.573154 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:27:26 crc kubenswrapper[4912]: I1203 01:27:26.220596 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" event={"ID":"d31b18ef-e64f-47ac-9aad-401795f9e3f0","Type":"ContainerStarted","Data":"6508d63444a001f3738c746a839dd42bc2384870e333b6310b11f7095d2a41ab"} Dec 03 01:27:26 crc kubenswrapper[4912]: I1203 01:27:26.221034 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" event={"ID":"d31b18ef-e64f-47ac-9aad-401795f9e3f0","Type":"ContainerStarted","Data":"676d53c0344e2c7f498f8f489a715047b3a27a1a87bf2cf97f724dc2dfdfed5d"} Dec 03 01:27:26 crc kubenswrapper[4912]: I1203 01:27:26.244876 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" podStartSLOduration=1.78521617 podStartE2EDuration="2.244859076s" podCreationTimestamp="2025-12-03 01:27:24 +0000 UTC" firstStartedPulling="2025-12-03 01:27:25.393449147 +0000 UTC m=+3831.035469717" lastFinishedPulling="2025-12-03 01:27:25.853092053 +0000 UTC m=+3831.495112623" observedRunningTime="2025-12-03 01:27:26.240684614 +0000 UTC m=+3831.882705184" watchObservedRunningTime="2025-12-03 01:27:26.244859076 +0000 UTC m=+3831.886879646" Dec 03 01:27:37 crc kubenswrapper[4912]: I1203 01:27:37.344415 4912 generic.go:334] "Generic (PLEG): container finished" podID="d31b18ef-e64f-47ac-9aad-401795f9e3f0" containerID="6508d63444a001f3738c746a839dd42bc2384870e333b6310b11f7095d2a41ab" exitCode=0 Dec 03 01:27:37 crc kubenswrapper[4912]: I1203 01:27:37.344497 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" event={"ID":"d31b18ef-e64f-47ac-9aad-401795f9e3f0","Type":"ContainerDied","Data":"6508d63444a001f3738c746a839dd42bc2384870e333b6310b11f7095d2a41ab"} Dec 03 01:27:37 crc kubenswrapper[4912]: I1203 01:27:37.572193 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:27:37 crc kubenswrapper[4912]: E1203 01:27:37.572490 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.010406 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.193386 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ceph\") pod \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.193648 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-inventory\") pod \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.193760 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xw7f\" (UniqueName: \"kubernetes.io/projected/d31b18ef-e64f-47ac-9aad-401795f9e3f0-kube-api-access-5xw7f\") pod \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.193793 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ssh-key\") pod \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\" (UID: \"d31b18ef-e64f-47ac-9aad-401795f9e3f0\") " Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.200049 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d31b18ef-e64f-47ac-9aad-401795f9e3f0-kube-api-access-5xw7f" (OuterVolumeSpecName: "kube-api-access-5xw7f") pod "d31b18ef-e64f-47ac-9aad-401795f9e3f0" (UID: "d31b18ef-e64f-47ac-9aad-401795f9e3f0"). InnerVolumeSpecName "kube-api-access-5xw7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.202965 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ceph" (OuterVolumeSpecName: "ceph") pod "d31b18ef-e64f-47ac-9aad-401795f9e3f0" (UID: "d31b18ef-e64f-47ac-9aad-401795f9e3f0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.222845 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d31b18ef-e64f-47ac-9aad-401795f9e3f0" (UID: "d31b18ef-e64f-47ac-9aad-401795f9e3f0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.230281 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-inventory" (OuterVolumeSpecName: "inventory") pod "d31b18ef-e64f-47ac-9aad-401795f9e3f0" (UID: "d31b18ef-e64f-47ac-9aad-401795f9e3f0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.296854 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xw7f\" (UniqueName: \"kubernetes.io/projected/d31b18ef-e64f-47ac-9aad-401795f9e3f0-kube-api-access-5xw7f\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.296902 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.296923 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.296939 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d31b18ef-e64f-47ac-9aad-401795f9e3f0-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.369204 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" event={"ID":"d31b18ef-e64f-47ac-9aad-401795f9e3f0","Type":"ContainerDied","Data":"676d53c0344e2c7f498f8f489a715047b3a27a1a87bf2cf97f724dc2dfdfed5d"} Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.369264 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="676d53c0344e2c7f498f8f489a715047b3a27a1a87bf2cf97f724dc2dfdfed5d" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.369275 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6kzw8" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.480318 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh"] Dec 03 01:27:39 crc kubenswrapper[4912]: E1203 01:27:39.480980 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d31b18ef-e64f-47ac-9aad-401795f9e3f0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.481030 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d31b18ef-e64f-47ac-9aad-401795f9e3f0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.481388 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d31b18ef-e64f-47ac-9aad-401795f9e3f0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.482421 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.493819 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.494319 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.494757 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.494857 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.494942 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.495128 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh"] Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.605379 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.605745 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.605887 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cwfz\" (UniqueName: \"kubernetes.io/projected/4d2c5206-f0c2-4df9-9a99-345471f00226-kube-api-access-8cwfz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.605930 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.709732 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cwfz\" (UniqueName: \"kubernetes.io/projected/4d2c5206-f0c2-4df9-9a99-345471f00226-kube-api-access-8cwfz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.709880 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.710073 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.710153 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.714802 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.718609 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.725807 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.736324 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cwfz\" (UniqueName: \"kubernetes.io/projected/4d2c5206-f0c2-4df9-9a99-345471f00226-kube-api-access-8cwfz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:39 crc kubenswrapper[4912]: I1203 01:27:39.814085 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:40 crc kubenswrapper[4912]: I1203 01:27:40.458830 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh"] Dec 03 01:27:41 crc kubenswrapper[4912]: I1203 01:27:41.393849 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" event={"ID":"4d2c5206-f0c2-4df9-9a99-345471f00226","Type":"ContainerStarted","Data":"32e529adba68ec53e2b3e39ab97df7a3db2905c7a92d9683b247734ea0f0107e"} Dec 03 01:27:41 crc kubenswrapper[4912]: I1203 01:27:41.394248 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" event={"ID":"4d2c5206-f0c2-4df9-9a99-345471f00226","Type":"ContainerStarted","Data":"d9598716c94de92e86429d6fc109770f6c86c515c0a5f5ed83b109c3c822eaf9"} Dec 03 01:27:41 crc kubenswrapper[4912]: I1203 01:27:41.425690 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" podStartSLOduration=1.937398881 podStartE2EDuration="2.425660697s" podCreationTimestamp="2025-12-03 01:27:39 +0000 UTC" firstStartedPulling="2025-12-03 01:27:40.458665722 +0000 UTC m=+3846.100686312" lastFinishedPulling="2025-12-03 01:27:40.946927548 +0000 UTC m=+3846.588948128" observedRunningTime="2025-12-03 01:27:41.418885735 +0000 UTC m=+3847.060906325" watchObservedRunningTime="2025-12-03 01:27:41.425660697 +0000 UTC m=+3847.067681277" Dec 03 01:27:48 crc kubenswrapper[4912]: I1203 01:27:48.571542 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:27:48 crc kubenswrapper[4912]: E1203 01:27:48.572598 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:27:56 crc kubenswrapper[4912]: I1203 01:27:56.564765 4912 generic.go:334] "Generic (PLEG): container finished" podID="4d2c5206-f0c2-4df9-9a99-345471f00226" containerID="32e529adba68ec53e2b3e39ab97df7a3db2905c7a92d9683b247734ea0f0107e" exitCode=0 Dec 03 01:27:56 crc kubenswrapper[4912]: I1203 01:27:56.564910 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" event={"ID":"4d2c5206-f0c2-4df9-9a99-345471f00226","Type":"ContainerDied","Data":"32e529adba68ec53e2b3e39ab97df7a3db2905c7a92d9683b247734ea0f0107e"} Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.108507 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.212679 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ceph\") pod \"4d2c5206-f0c2-4df9-9a99-345471f00226\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.212757 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-inventory\") pod \"4d2c5206-f0c2-4df9-9a99-345471f00226\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.212788 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ssh-key\") pod \"4d2c5206-f0c2-4df9-9a99-345471f00226\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.212822 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cwfz\" (UniqueName: \"kubernetes.io/projected/4d2c5206-f0c2-4df9-9a99-345471f00226-kube-api-access-8cwfz\") pod \"4d2c5206-f0c2-4df9-9a99-345471f00226\" (UID: \"4d2c5206-f0c2-4df9-9a99-345471f00226\") " Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.219414 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ceph" (OuterVolumeSpecName: "ceph") pod "4d2c5206-f0c2-4df9-9a99-345471f00226" (UID: "4d2c5206-f0c2-4df9-9a99-345471f00226"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.221092 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d2c5206-f0c2-4df9-9a99-345471f00226-kube-api-access-8cwfz" (OuterVolumeSpecName: "kube-api-access-8cwfz") pod "4d2c5206-f0c2-4df9-9a99-345471f00226" (UID: "4d2c5206-f0c2-4df9-9a99-345471f00226"). InnerVolumeSpecName "kube-api-access-8cwfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.242734 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-inventory" (OuterVolumeSpecName: "inventory") pod "4d2c5206-f0c2-4df9-9a99-345471f00226" (UID: "4d2c5206-f0c2-4df9-9a99-345471f00226"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.249981 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d2c5206-f0c2-4df9-9a99-345471f00226" (UID: "4d2c5206-f0c2-4df9-9a99-345471f00226"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.315916 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.315959 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.315972 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cwfz\" (UniqueName: \"kubernetes.io/projected/4d2c5206-f0c2-4df9-9a99-345471f00226-kube-api-access-8cwfz\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.315984 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4d2c5206-f0c2-4df9-9a99-345471f00226-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.603157 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" event={"ID":"4d2c5206-f0c2-4df9-9a99-345471f00226","Type":"ContainerDied","Data":"d9598716c94de92e86429d6fc109770f6c86c515c0a5f5ed83b109c3c822eaf9"} Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.603194 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9598716c94de92e86429d6fc109770f6c86c515c0a5f5ed83b109c3c822eaf9" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.603201 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.687752 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98"] Dec 03 01:27:58 crc kubenswrapper[4912]: E1203 01:27:58.688274 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d2c5206-f0c2-4df9-9a99-345471f00226" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.688295 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d2c5206-f0c2-4df9-9a99-345471f00226" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.688622 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d2c5206-f0c2-4df9-9a99-345471f00226" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.689519 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.692574 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.693215 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.693519 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.693665 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.693829 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.694537 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.696030 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.703356 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98"] Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.704992 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.704997 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.706766 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.826853 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.826901 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.826945 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.826975 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827188 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827235 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827268 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827287 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827326 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827347 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827374 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827404 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827448 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cvj6\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-kube-api-access-8cvj6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827468 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827510 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827556 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.827583 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929548 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929610 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929642 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929670 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929719 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929740 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929775 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929801 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929830 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929847 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929891 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929915 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929943 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929965 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.929991 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cvj6\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-kube-api-access-8cvj6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.930011 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.930042 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.940370 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.947334 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.947574 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.947869 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.948300 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.951947 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.956459 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.958163 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.958297 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.958309 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.958480 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.958506 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.958539 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.959029 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.959629 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cvj6\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-kube-api-access-8cvj6\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.964870 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:58 crc kubenswrapper[4912]: I1203 01:27:58.970966 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-ssn98\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:59 crc kubenswrapper[4912]: I1203 01:27:59.037234 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:27:59 crc kubenswrapper[4912]: I1203 01:27:59.623933 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" event={"ID":"d2d7b572-6166-47b3-9346-17f881d869a7","Type":"ContainerStarted","Data":"7a5b00165a406dcbef8be0626e1ceab48d3795b7cc2cd6c3ad4c2877b7a27262"} Dec 03 01:27:59 crc kubenswrapper[4912]: I1203 01:27:59.647063 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98"] Dec 03 01:28:00 crc kubenswrapper[4912]: I1203 01:28:00.637007 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" event={"ID":"d2d7b572-6166-47b3-9346-17f881d869a7","Type":"ContainerStarted","Data":"f8b2cdb76d4f44880396cdbf014867ceef065599f625d6c436c97603f36ac9bf"} Dec 03 01:28:00 crc kubenswrapper[4912]: I1203 01:28:00.667603 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" podStartSLOduration=2.161328037 podStartE2EDuration="2.667581525s" podCreationTimestamp="2025-12-03 01:27:58 +0000 UTC" firstStartedPulling="2025-12-03 01:27:59.596261588 +0000 UTC m=+3865.238282188" lastFinishedPulling="2025-12-03 01:28:00.102515116 +0000 UTC m=+3865.744535676" observedRunningTime="2025-12-03 01:28:00.658409538 +0000 UTC m=+3866.300430128" watchObservedRunningTime="2025-12-03 01:28:00.667581525 +0000 UTC m=+3866.309602105" Dec 03 01:28:02 crc kubenswrapper[4912]: I1203 01:28:02.573099 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:28:02 crc kubenswrapper[4912]: E1203 01:28:02.573946 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:28:17 crc kubenswrapper[4912]: I1203 01:28:17.573235 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:28:17 crc kubenswrapper[4912]: E1203 01:28:17.576223 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:28:28 crc kubenswrapper[4912]: I1203 01:28:28.572362 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:28:28 crc kubenswrapper[4912]: E1203 01:28:28.573548 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:28:42 crc kubenswrapper[4912]: I1203 01:28:42.575528 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:28:42 crc kubenswrapper[4912]: E1203 01:28:42.576512 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:28:53 crc kubenswrapper[4912]: I1203 01:28:53.571980 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:28:53 crc kubenswrapper[4912]: E1203 01:28:53.572676 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:29:06 crc kubenswrapper[4912]: I1203 01:29:06.572283 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:29:06 crc kubenswrapper[4912]: E1203 01:29:06.573081 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:29:19 crc kubenswrapper[4912]: I1203 01:29:19.617602 4912 generic.go:334] "Generic (PLEG): container finished" podID="d2d7b572-6166-47b3-9346-17f881d869a7" containerID="f8b2cdb76d4f44880396cdbf014867ceef065599f625d6c436c97603f36ac9bf" exitCode=0 Dec 03 01:29:19 crc kubenswrapper[4912]: I1203 01:29:19.617954 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" event={"ID":"d2d7b572-6166-47b3-9346-17f881d869a7","Type":"ContainerDied","Data":"f8b2cdb76d4f44880396cdbf014867ceef065599f625d6c436c97603f36ac9bf"} Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.136075 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.180821 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cvj6\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-kube-api-access-8cvj6\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.180945 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-combined-ca-bundle\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.180989 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-bootstrap-combined-ca-bundle\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181033 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ceph\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181070 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181106 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181201 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-inventory\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181238 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-power-monitoring-combined-ca-bundle\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181283 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181352 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-repo-setup-combined-ca-bundle\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181524 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ssh-key\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181595 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-nova-combined-ca-bundle\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181633 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-neutron-metadata-combined-ca-bundle\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181678 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ovn-combined-ca-bundle\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181757 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181797 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-libvirt-combined-ca-bundle\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.181841 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"d2d7b572-6166-47b3-9346-17f881d869a7\" (UID: \"d2d7b572-6166-47b3-9346-17f881d869a7\") " Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.190662 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.191154 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.192277 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.193268 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.193598 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.196155 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.196774 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-kube-api-access-8cvj6" (OuterVolumeSpecName: "kube-api-access-8cvj6") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "kube-api-access-8cvj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.197884 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.199753 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.203182 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.203270 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.208758 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.209858 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.210787 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.216129 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ceph" (OuterVolumeSpecName: "ceph") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.222798 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-inventory" (OuterVolumeSpecName: "inventory") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.229642 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2d7b572-6166-47b3-9346-17f881d869a7" (UID: "d2d7b572-6166-47b3-9346-17f881d869a7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284765 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cvj6\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-kube-api-access-8cvj6\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284800 4912 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284810 4912 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284818 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284829 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284841 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284851 4912 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284861 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284872 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284882 4912 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284890 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284898 4912 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284906 4912 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284914 4912 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284923 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284931 4912 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d7b572-6166-47b3-9346-17f881d869a7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.284942 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d2d7b572-6166-47b3-9346-17f881d869a7-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.573822 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.642458 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" event={"ID":"d2d7b572-6166-47b3-9346-17f881d869a7","Type":"ContainerDied","Data":"7a5b00165a406dcbef8be0626e1ceab48d3795b7cc2cd6c3ad4c2877b7a27262"} Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.642500 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a5b00165a406dcbef8be0626e1ceab48d3795b7cc2cd6c3ad4c2877b7a27262" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.642584 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-ssn98" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.857344 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv"] Dec 03 01:29:21 crc kubenswrapper[4912]: E1203 01:29:21.858262 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d7b572-6166-47b3-9346-17f881d869a7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.858297 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d7b572-6166-47b3-9346-17f881d869a7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.858646 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d7b572-6166-47b3-9346-17f881d869a7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.860068 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.861807 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.861917 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.862465 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.862693 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.863549 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.877360 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv"] Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.903394 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcn97\" (UniqueName: \"kubernetes.io/projected/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-kube-api-access-bcn97\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.903441 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.903485 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:21 crc kubenswrapper[4912]: I1203 01:29:21.903520 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.005079 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcn97\" (UniqueName: \"kubernetes.io/projected/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-kube-api-access-bcn97\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.005124 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.005143 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.005175 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.489327 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.489372 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.489801 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.490373 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcn97\" (UniqueName: \"kubernetes.io/projected/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-kube-api-access-bcn97\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.653674 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"f5ddc05f695c6bfb74ae5cda3f4bce456d3a1a95d24096442007066627b0d1c3"} Dec 03 01:29:22 crc kubenswrapper[4912]: I1203 01:29:22.779761 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:23 crc kubenswrapper[4912]: I1203 01:29:23.337777 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv"] Dec 03 01:29:23 crc kubenswrapper[4912]: W1203 01:29:23.345132 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb529c480_6fb1_4a0b_a915_1aca2f6e85bd.slice/crio-1b0d4d896d618aa8f1ac2c93f39cae4af74141972a3b09d2d44b615d1a3ef25c WatchSource:0}: Error finding container 1b0d4d896d618aa8f1ac2c93f39cae4af74141972a3b09d2d44b615d1a3ef25c: Status 404 returned error can't find the container with id 1b0d4d896d618aa8f1ac2c93f39cae4af74141972a3b09d2d44b615d1a3ef25c Dec 03 01:29:23 crc kubenswrapper[4912]: I1203 01:29:23.667129 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" event={"ID":"b529c480-6fb1-4a0b-a915-1aca2f6e85bd","Type":"ContainerStarted","Data":"1b0d4d896d618aa8f1ac2c93f39cae4af74141972a3b09d2d44b615d1a3ef25c"} Dec 03 01:29:25 crc kubenswrapper[4912]: I1203 01:29:25.694629 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" event={"ID":"b529c480-6fb1-4a0b-a915-1aca2f6e85bd","Type":"ContainerStarted","Data":"867aa54f15ff75de8b6eee12c2ab6f8078fea19de6e4fe4b971f27b88625fcb9"} Dec 03 01:29:25 crc kubenswrapper[4912]: I1203 01:29:25.724005 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" podStartSLOduration=3.873116476 podStartE2EDuration="4.723975919s" podCreationTimestamp="2025-12-03 01:29:21 +0000 UTC" firstStartedPulling="2025-12-03 01:29:23.347764233 +0000 UTC m=+3948.989784803" lastFinishedPulling="2025-12-03 01:29:24.198623656 +0000 UTC m=+3949.840644246" observedRunningTime="2025-12-03 01:29:25.718085173 +0000 UTC m=+3951.360105763" watchObservedRunningTime="2025-12-03 01:29:25.723975919 +0000 UTC m=+3951.365996529" Dec 03 01:29:33 crc kubenswrapper[4912]: I1203 01:29:33.794369 4912 generic.go:334] "Generic (PLEG): container finished" podID="b529c480-6fb1-4a0b-a915-1aca2f6e85bd" containerID="867aa54f15ff75de8b6eee12c2ab6f8078fea19de6e4fe4b971f27b88625fcb9" exitCode=0 Dec 03 01:29:33 crc kubenswrapper[4912]: I1203 01:29:33.794584 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" event={"ID":"b529c480-6fb1-4a0b-a915-1aca2f6e85bd","Type":"ContainerDied","Data":"867aa54f15ff75de8b6eee12c2ab6f8078fea19de6e4fe4b971f27b88625fcb9"} Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.028502 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.143697 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-inventory\") pod \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.144476 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ceph\") pod \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.144579 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcn97\" (UniqueName: \"kubernetes.io/projected/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-kube-api-access-bcn97\") pod \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.144725 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ssh-key\") pod \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\" (UID: \"b529c480-6fb1-4a0b-a915-1aca2f6e85bd\") " Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.151464 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ceph" (OuterVolumeSpecName: "ceph") pod "b529c480-6fb1-4a0b-a915-1aca2f6e85bd" (UID: "b529c480-6fb1-4a0b-a915-1aca2f6e85bd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.151753 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-kube-api-access-bcn97" (OuterVolumeSpecName: "kube-api-access-bcn97") pod "b529c480-6fb1-4a0b-a915-1aca2f6e85bd" (UID: "b529c480-6fb1-4a0b-a915-1aca2f6e85bd"). InnerVolumeSpecName "kube-api-access-bcn97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.177420 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-inventory" (OuterVolumeSpecName: "inventory") pod "b529c480-6fb1-4a0b-a915-1aca2f6e85bd" (UID: "b529c480-6fb1-4a0b-a915-1aca2f6e85bd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.197786 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b529c480-6fb1-4a0b-a915-1aca2f6e85bd" (UID: "b529c480-6fb1-4a0b-a915-1aca2f6e85bd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.247495 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.247580 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.247597 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.247610 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcn97\" (UniqueName: \"kubernetes.io/projected/b529c480-6fb1-4a0b-a915-1aca2f6e85bd-kube-api-access-bcn97\") on node \"crc\" DevicePath \"\"" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.826774 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" event={"ID":"b529c480-6fb1-4a0b-a915-1aca2f6e85bd","Type":"ContainerDied","Data":"1b0d4d896d618aa8f1ac2c93f39cae4af74141972a3b09d2d44b615d1a3ef25c"} Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.826818 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv" Dec 03 01:29:36 crc kubenswrapper[4912]: I1203 01:29:36.826821 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b0d4d896d618aa8f1ac2c93f39cae4af74141972a3b09d2d44b615d1a3ef25c" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.161294 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj"] Dec 03 01:29:37 crc kubenswrapper[4912]: E1203 01:29:37.161988 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b529c480-6fb1-4a0b-a915-1aca2f6e85bd" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.162013 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b529c480-6fb1-4a0b-a915-1aca2f6e85bd" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.162357 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="b529c480-6fb1-4a0b-a915-1aca2f6e85bd" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.163565 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.167193 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.167370 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.167405 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.167626 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.167766 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.167961 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.177054 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj"] Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.268615 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/04c823be-fed9-4ffd-b41b-30115e049634-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.268680 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.268713 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5wj4\" (UniqueName: \"kubernetes.io/projected/04c823be-fed9-4ffd-b41b-30115e049634-kube-api-access-x5wj4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.268739 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.268806 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.268893 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.370386 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/04c823be-fed9-4ffd-b41b-30115e049634-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.370445 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.370473 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5wj4\" (UniqueName: \"kubernetes.io/projected/04c823be-fed9-4ffd-b41b-30115e049634-kube-api-access-x5wj4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.370495 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.370551 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.370616 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.372478 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/04c823be-fed9-4ffd-b41b-30115e049634-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.375703 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.390922 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.398026 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.399976 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.400877 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5wj4\" (UniqueName: \"kubernetes.io/projected/04c823be-fed9-4ffd-b41b-30115e049634-kube-api-access-x5wj4\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-49qvj\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:37 crc kubenswrapper[4912]: I1203 01:29:37.504771 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:29:38 crc kubenswrapper[4912]: I1203 01:29:38.099131 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj"] Dec 03 01:29:38 crc kubenswrapper[4912]: W1203 01:29:38.393023 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04c823be_fed9_4ffd_b41b_30115e049634.slice/crio-a1129f70beddb4c586af3a16feb3ee95da7125a2f41229b9547c96009ab9e5bf WatchSource:0}: Error finding container a1129f70beddb4c586af3a16feb3ee95da7125a2f41229b9547c96009ab9e5bf: Status 404 returned error can't find the container with id a1129f70beddb4c586af3a16feb3ee95da7125a2f41229b9547c96009ab9e5bf Dec 03 01:29:38 crc kubenswrapper[4912]: I1203 01:29:38.858147 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" event={"ID":"04c823be-fed9-4ffd-b41b-30115e049634","Type":"ContainerStarted","Data":"a1129f70beddb4c586af3a16feb3ee95da7125a2f41229b9547c96009ab9e5bf"} Dec 03 01:29:39 crc kubenswrapper[4912]: I1203 01:29:39.869056 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" event={"ID":"04c823be-fed9-4ffd-b41b-30115e049634","Type":"ContainerStarted","Data":"475e6fb30807bbf641bd02d0412ae979391398bb0de9460cc7a5446ba2a3eeec"} Dec 03 01:29:39 crc kubenswrapper[4912]: I1203 01:29:39.902038 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" podStartSLOduration=2.47298442 podStartE2EDuration="2.902011436s" podCreationTimestamp="2025-12-03 01:29:37 +0000 UTC" firstStartedPulling="2025-12-03 01:29:38.396029698 +0000 UTC m=+3964.038050258" lastFinishedPulling="2025-12-03 01:29:38.825056674 +0000 UTC m=+3964.467077274" observedRunningTime="2025-12-03 01:29:39.884983914 +0000 UTC m=+3965.527004495" watchObservedRunningTime="2025-12-03 01:29:39.902011436 +0000 UTC m=+3965.544032026" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.186896 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w"] Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.189005 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.191487 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.199366 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.203329 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w"] Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.304227 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01fab78-4545-4c05-a3c2-ed7d32cc222e-secret-volume\") pod \"collect-profiles-29412090-hgj8w\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.304349 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01fab78-4545-4c05-a3c2-ed7d32cc222e-config-volume\") pod \"collect-profiles-29412090-hgj8w\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.304572 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c6hg\" (UniqueName: \"kubernetes.io/projected/c01fab78-4545-4c05-a3c2-ed7d32cc222e-kube-api-access-7c6hg\") pod \"collect-profiles-29412090-hgj8w\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.407719 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01fab78-4545-4c05-a3c2-ed7d32cc222e-secret-volume\") pod \"collect-profiles-29412090-hgj8w\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.407873 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01fab78-4545-4c05-a3c2-ed7d32cc222e-config-volume\") pod \"collect-profiles-29412090-hgj8w\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.407984 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c6hg\" (UniqueName: \"kubernetes.io/projected/c01fab78-4545-4c05-a3c2-ed7d32cc222e-kube-api-access-7c6hg\") pod \"collect-profiles-29412090-hgj8w\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.409685 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01fab78-4545-4c05-a3c2-ed7d32cc222e-config-volume\") pod \"collect-profiles-29412090-hgj8w\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.417393 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01fab78-4545-4c05-a3c2-ed7d32cc222e-secret-volume\") pod \"collect-profiles-29412090-hgj8w\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.450557 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c6hg\" (UniqueName: \"kubernetes.io/projected/c01fab78-4545-4c05-a3c2-ed7d32cc222e-kube-api-access-7c6hg\") pod \"collect-profiles-29412090-hgj8w\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:00 crc kubenswrapper[4912]: I1203 01:30:00.521697 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:01 crc kubenswrapper[4912]: I1203 01:30:01.070556 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w"] Dec 03 01:30:01 crc kubenswrapper[4912]: I1203 01:30:01.111947 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" event={"ID":"c01fab78-4545-4c05-a3c2-ed7d32cc222e","Type":"ContainerStarted","Data":"08fd99b03c21890b3da49ed89865f95207fac1aa450a72d51451ba05c3bb590c"} Dec 03 01:30:02 crc kubenswrapper[4912]: I1203 01:30:02.131899 4912 generic.go:334] "Generic (PLEG): container finished" podID="c01fab78-4545-4c05-a3c2-ed7d32cc222e" containerID="969f718bec11c2f1e6ad4f4790843029de5ae39182502f83370ee93ed37f4cf8" exitCode=0 Dec 03 01:30:02 crc kubenswrapper[4912]: I1203 01:30:02.131973 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" event={"ID":"c01fab78-4545-4c05-a3c2-ed7d32cc222e","Type":"ContainerDied","Data":"969f718bec11c2f1e6ad4f4790843029de5ae39182502f83370ee93ed37f4cf8"} Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.663685 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.815162 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01fab78-4545-4c05-a3c2-ed7d32cc222e-secret-volume\") pod \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.815338 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c6hg\" (UniqueName: \"kubernetes.io/projected/c01fab78-4545-4c05-a3c2-ed7d32cc222e-kube-api-access-7c6hg\") pod \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.815463 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01fab78-4545-4c05-a3c2-ed7d32cc222e-config-volume\") pod \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\" (UID: \"c01fab78-4545-4c05-a3c2-ed7d32cc222e\") " Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.816714 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c01fab78-4545-4c05-a3c2-ed7d32cc222e-config-volume" (OuterVolumeSpecName: "config-volume") pod "c01fab78-4545-4c05-a3c2-ed7d32cc222e" (UID: "c01fab78-4545-4c05-a3c2-ed7d32cc222e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.822052 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c01fab78-4545-4c05-a3c2-ed7d32cc222e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c01fab78-4545-4c05-a3c2-ed7d32cc222e" (UID: "c01fab78-4545-4c05-a3c2-ed7d32cc222e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.823000 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c01fab78-4545-4c05-a3c2-ed7d32cc222e-kube-api-access-7c6hg" (OuterVolumeSpecName: "kube-api-access-7c6hg") pod "c01fab78-4545-4c05-a3c2-ed7d32cc222e" (UID: "c01fab78-4545-4c05-a3c2-ed7d32cc222e"). InnerVolumeSpecName "kube-api-access-7c6hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.918753 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01fab78-4545-4c05-a3c2-ed7d32cc222e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.918821 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c6hg\" (UniqueName: \"kubernetes.io/projected/c01fab78-4545-4c05-a3c2-ed7d32cc222e-kube-api-access-7c6hg\") on node \"crc\" DevicePath \"\"" Dec 03 01:30:03 crc kubenswrapper[4912]: I1203 01:30:03.918851 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01fab78-4545-4c05-a3c2-ed7d32cc222e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 01:30:04 crc kubenswrapper[4912]: I1203 01:30:04.158948 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" event={"ID":"c01fab78-4545-4c05-a3c2-ed7d32cc222e","Type":"ContainerDied","Data":"08fd99b03c21890b3da49ed89865f95207fac1aa450a72d51451ba05c3bb590c"} Dec 03 01:30:04 crc kubenswrapper[4912]: I1203 01:30:04.159352 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08fd99b03c21890b3da49ed89865f95207fac1aa450a72d51451ba05c3bb590c" Dec 03 01:30:04 crc kubenswrapper[4912]: I1203 01:30:04.159014 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w" Dec 03 01:30:04 crc kubenswrapper[4912]: I1203 01:30:04.761023 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b"] Dec 03 01:30:04 crc kubenswrapper[4912]: I1203 01:30:04.771363 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412045-hh55b"] Dec 03 01:30:06 crc kubenswrapper[4912]: I1203 01:30:06.604741 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5be2251-7d9d-4360-8e88-d71bc8458cf8" path="/var/lib/kubelet/pods/a5be2251-7d9d-4360-8e88-d71bc8458cf8/volumes" Dec 03 01:30:25 crc kubenswrapper[4912]: I1203 01:30:25.348424 4912 scope.go:117] "RemoveContainer" containerID="5bdc7bf2168fd20ce5de41930c506e045710fcf1a64f7e19bee6006c69e5b0a3" Dec 03 01:31:19 crc kubenswrapper[4912]: I1203 01:31:19.130048 4912 generic.go:334] "Generic (PLEG): container finished" podID="04c823be-fed9-4ffd-b41b-30115e049634" containerID="475e6fb30807bbf641bd02d0412ae979391398bb0de9460cc7a5446ba2a3eeec" exitCode=0 Dec 03 01:31:19 crc kubenswrapper[4912]: I1203 01:31:19.130135 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" event={"ID":"04c823be-fed9-4ffd-b41b-30115e049634","Type":"ContainerDied","Data":"475e6fb30807bbf641bd02d0412ae979391398bb0de9460cc7a5446ba2a3eeec"} Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.674401 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.767479 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/04c823be-fed9-4ffd-b41b-30115e049634-ovncontroller-config-0\") pod \"04c823be-fed9-4ffd-b41b-30115e049634\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.767878 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5wj4\" (UniqueName: \"kubernetes.io/projected/04c823be-fed9-4ffd-b41b-30115e049634-kube-api-access-x5wj4\") pod \"04c823be-fed9-4ffd-b41b-30115e049634\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.767966 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ovn-combined-ca-bundle\") pod \"04c823be-fed9-4ffd-b41b-30115e049634\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.768001 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ssh-key\") pod \"04c823be-fed9-4ffd-b41b-30115e049634\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.768025 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-inventory\") pod \"04c823be-fed9-4ffd-b41b-30115e049634\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.768143 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ceph\") pod \"04c823be-fed9-4ffd-b41b-30115e049634\" (UID: \"04c823be-fed9-4ffd-b41b-30115e049634\") " Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.774601 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04c823be-fed9-4ffd-b41b-30115e049634-kube-api-access-x5wj4" (OuterVolumeSpecName: "kube-api-access-x5wj4") pod "04c823be-fed9-4ffd-b41b-30115e049634" (UID: "04c823be-fed9-4ffd-b41b-30115e049634"). InnerVolumeSpecName "kube-api-access-x5wj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.776136 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "04c823be-fed9-4ffd-b41b-30115e049634" (UID: "04c823be-fed9-4ffd-b41b-30115e049634"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.780016 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ceph" (OuterVolumeSpecName: "ceph") pod "04c823be-fed9-4ffd-b41b-30115e049634" (UID: "04c823be-fed9-4ffd-b41b-30115e049634"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.809545 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "04c823be-fed9-4ffd-b41b-30115e049634" (UID: "04c823be-fed9-4ffd-b41b-30115e049634"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.813371 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-inventory" (OuterVolumeSpecName: "inventory") pod "04c823be-fed9-4ffd-b41b-30115e049634" (UID: "04c823be-fed9-4ffd-b41b-30115e049634"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.824738 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04c823be-fed9-4ffd-b41b-30115e049634-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "04c823be-fed9-4ffd-b41b-30115e049634" (UID: "04c823be-fed9-4ffd-b41b-30115e049634"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.871419 4912 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/04c823be-fed9-4ffd-b41b-30115e049634-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.871499 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5wj4\" (UniqueName: \"kubernetes.io/projected/04c823be-fed9-4ffd-b41b-30115e049634-kube-api-access-x5wj4\") on node \"crc\" DevicePath \"\"" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.871529 4912 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.871558 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.871587 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:31:20 crc kubenswrapper[4912]: I1203 01:31:20.871760 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/04c823be-fed9-4ffd-b41b-30115e049634-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.166233 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" event={"ID":"04c823be-fed9-4ffd-b41b-30115e049634","Type":"ContainerDied","Data":"a1129f70beddb4c586af3a16feb3ee95da7125a2f41229b9547c96009ab9e5bf"} Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.166280 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1129f70beddb4c586af3a16feb3ee95da7125a2f41229b9547c96009ab9e5bf" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.166318 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-49qvj" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.461299 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747"] Dec 03 01:31:21 crc kubenswrapper[4912]: E1203 01:31:21.461968 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01fab78-4545-4c05-a3c2-ed7d32cc222e" containerName="collect-profiles" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.461996 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01fab78-4545-4c05-a3c2-ed7d32cc222e" containerName="collect-profiles" Dec 03 01:31:21 crc kubenswrapper[4912]: E1203 01:31:21.462105 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04c823be-fed9-4ffd-b41b-30115e049634" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.462123 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="04c823be-fed9-4ffd-b41b-30115e049634" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.462654 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01fab78-4545-4c05-a3c2-ed7d32cc222e" containerName="collect-profiles" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.462725 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="04c823be-fed9-4ffd-b41b-30115e049634" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.463947 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.468772 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.468999 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.469003 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.469058 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.469947 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.469082 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.469086 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.484484 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747"] Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.627156 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.627614 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.627816 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.627912 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.628016 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.628103 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.628210 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blq4n\" (UniqueName: \"kubernetes.io/projected/748f1896-9a69-406c-bc59-0662de1da921-kube-api-access-blq4n\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.729855 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.729925 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.729974 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.730009 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.730055 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blq4n\" (UniqueName: \"kubernetes.io/projected/748f1896-9a69-406c-bc59-0662de1da921-kube-api-access-blq4n\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.730155 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.730189 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.735994 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.736176 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.737782 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.737865 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.738068 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.739426 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.750718 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blq4n\" (UniqueName: \"kubernetes.io/projected/748f1896-9a69-406c-bc59-0662de1da921-kube-api-access-blq4n\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:21 crc kubenswrapper[4912]: I1203 01:31:21.806262 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:31:22 crc kubenswrapper[4912]: I1203 01:31:22.495574 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747"] Dec 03 01:31:23 crc kubenswrapper[4912]: I1203 01:31:23.192554 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" event={"ID":"748f1896-9a69-406c-bc59-0662de1da921","Type":"ContainerStarted","Data":"e97d456edb3e067faeece2f9cefc0a2ebe4b0c2f7167c1d4890accbed233a737"} Dec 03 01:31:24 crc kubenswrapper[4912]: I1203 01:31:24.207338 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" event={"ID":"748f1896-9a69-406c-bc59-0662de1da921","Type":"ContainerStarted","Data":"01923f143ba28d7a42568dd27e12dd433c1c73d48a2ef2308ed9f921e70c4d70"} Dec 03 01:31:24 crc kubenswrapper[4912]: I1203 01:31:24.253413 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" podStartSLOduration=2.70402729 podStartE2EDuration="3.25338334s" podCreationTimestamp="2025-12-03 01:31:21 +0000 UTC" firstStartedPulling="2025-12-03 01:31:22.489650061 +0000 UTC m=+4068.131670651" lastFinishedPulling="2025-12-03 01:31:23.039006101 +0000 UTC m=+4068.681026701" observedRunningTime="2025-12-03 01:31:24.239364078 +0000 UTC m=+4069.881384648" watchObservedRunningTime="2025-12-03 01:31:24.25338334 +0000 UTC m=+4069.895403940" Dec 03 01:31:48 crc kubenswrapper[4912]: I1203 01:31:48.077518 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:31:48 crc kubenswrapper[4912]: I1203 01:31:48.078044 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:32:18 crc kubenswrapper[4912]: I1203 01:32:18.077747 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:32:18 crc kubenswrapper[4912]: I1203 01:32:18.078335 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.635518 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6tw66"] Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.639976 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.658852 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tw66"] Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.722914 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-utilities\") pod \"redhat-marketplace-6tw66\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.723211 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhkzj\" (UniqueName: \"kubernetes.io/projected/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-kube-api-access-fhkzj\") pod \"redhat-marketplace-6tw66\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.723290 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-catalog-content\") pod \"redhat-marketplace-6tw66\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.825709 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhkzj\" (UniqueName: \"kubernetes.io/projected/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-kube-api-access-fhkzj\") pod \"redhat-marketplace-6tw66\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.825797 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-catalog-content\") pod \"redhat-marketplace-6tw66\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.825971 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-utilities\") pod \"redhat-marketplace-6tw66\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.826651 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-catalog-content\") pod \"redhat-marketplace-6tw66\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.826835 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-utilities\") pod \"redhat-marketplace-6tw66\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.859012 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhkzj\" (UniqueName: \"kubernetes.io/projected/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-kube-api-access-fhkzj\") pod \"redhat-marketplace-6tw66\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:47 crc kubenswrapper[4912]: I1203 01:32:47.975525 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:48 crc kubenswrapper[4912]: I1203 01:32:48.081271 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:32:48 crc kubenswrapper[4912]: I1203 01:32:48.081321 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:32:48 crc kubenswrapper[4912]: I1203 01:32:48.081363 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:32:48 crc kubenswrapper[4912]: I1203 01:32:48.082151 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5ddc05f695c6bfb74ae5cda3f4bce456d3a1a95d24096442007066627b0d1c3"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:32:48 crc kubenswrapper[4912]: I1203 01:32:48.082203 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://f5ddc05f695c6bfb74ae5cda3f4bce456d3a1a95d24096442007066627b0d1c3" gracePeriod=600 Dec 03 01:32:48 crc kubenswrapper[4912]: I1203 01:32:48.302108 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="f5ddc05f695c6bfb74ae5cda3f4bce456d3a1a95d24096442007066627b0d1c3" exitCode=0 Dec 03 01:32:48 crc kubenswrapper[4912]: I1203 01:32:48.302195 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"f5ddc05f695c6bfb74ae5cda3f4bce456d3a1a95d24096442007066627b0d1c3"} Dec 03 01:32:48 crc kubenswrapper[4912]: I1203 01:32:48.302665 4912 scope.go:117] "RemoveContainer" containerID="b8c7a6e2900f5b3f583017f6ac2085f22f7ab6500a7b82521f8fb078d4e376e2" Dec 03 01:32:48 crc kubenswrapper[4912]: I1203 01:32:48.541502 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tw66"] Dec 03 01:32:49 crc kubenswrapper[4912]: I1203 01:32:49.323595 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00"} Dec 03 01:32:49 crc kubenswrapper[4912]: I1203 01:32:49.327692 4912 generic.go:334] "Generic (PLEG): container finished" podID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerID="ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446" exitCode=0 Dec 03 01:32:49 crc kubenswrapper[4912]: I1203 01:32:49.327758 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tw66" event={"ID":"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0","Type":"ContainerDied","Data":"ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446"} Dec 03 01:32:49 crc kubenswrapper[4912]: I1203 01:32:49.327851 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tw66" event={"ID":"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0","Type":"ContainerStarted","Data":"eeaa9a1bb1f7e843e5d247228522154f98a6a532a247f2bf2e914de11ea018cb"} Dec 03 01:32:49 crc kubenswrapper[4912]: I1203 01:32:49.331033 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:32:50 crc kubenswrapper[4912]: I1203 01:32:50.342054 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tw66" event={"ID":"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0","Type":"ContainerStarted","Data":"a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e"} Dec 03 01:32:51 crc kubenswrapper[4912]: I1203 01:32:51.355524 4912 generic.go:334] "Generic (PLEG): container finished" podID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerID="a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e" exitCode=0 Dec 03 01:32:51 crc kubenswrapper[4912]: I1203 01:32:51.355597 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tw66" event={"ID":"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0","Type":"ContainerDied","Data":"a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e"} Dec 03 01:32:52 crc kubenswrapper[4912]: I1203 01:32:52.368721 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tw66" event={"ID":"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0","Type":"ContainerStarted","Data":"71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f"} Dec 03 01:32:52 crc kubenswrapper[4912]: I1203 01:32:52.389511 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6tw66" podStartSLOduration=2.921999826 podStartE2EDuration="5.389490958s" podCreationTimestamp="2025-12-03 01:32:47 +0000 UTC" firstStartedPulling="2025-12-03 01:32:49.330809399 +0000 UTC m=+4154.972829959" lastFinishedPulling="2025-12-03 01:32:51.798300511 +0000 UTC m=+4157.440321091" observedRunningTime="2025-12-03 01:32:52.387232712 +0000 UTC m=+4158.029253302" watchObservedRunningTime="2025-12-03 01:32:52.389490958 +0000 UTC m=+4158.031511528" Dec 03 01:32:57 crc kubenswrapper[4912]: I1203 01:32:57.976419 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:57 crc kubenswrapper[4912]: I1203 01:32:57.976957 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:58 crc kubenswrapper[4912]: I1203 01:32:58.047781 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:58 crc kubenswrapper[4912]: I1203 01:32:58.536853 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:32:58 crc kubenswrapper[4912]: I1203 01:32:58.605231 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tw66"] Dec 03 01:32:59 crc kubenswrapper[4912]: I1203 01:32:59.457811 4912 generic.go:334] "Generic (PLEG): container finished" podID="748f1896-9a69-406c-bc59-0662de1da921" containerID="01923f143ba28d7a42568dd27e12dd433c1c73d48a2ef2308ed9f921e70c4d70" exitCode=0 Dec 03 01:32:59 crc kubenswrapper[4912]: I1203 01:32:59.457927 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" event={"ID":"748f1896-9a69-406c-bc59-0662de1da921","Type":"ContainerDied","Data":"01923f143ba28d7a42568dd27e12dd433c1c73d48a2ef2308ed9f921e70c4d70"} Dec 03 01:33:00 crc kubenswrapper[4912]: I1203 01:33:00.469349 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6tw66" podUID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerName="registry-server" containerID="cri-o://71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f" gracePeriod=2 Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.100779 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.107156 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128311 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blq4n\" (UniqueName: \"kubernetes.io/projected/748f1896-9a69-406c-bc59-0662de1da921-kube-api-access-blq4n\") pod \"748f1896-9a69-406c-bc59-0662de1da921\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128390 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-nova-metadata-neutron-config-0\") pod \"748f1896-9a69-406c-bc59-0662de1da921\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128423 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ceph\") pod \"748f1896-9a69-406c-bc59-0662de1da921\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128503 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-metadata-combined-ca-bundle\") pod \"748f1896-9a69-406c-bc59-0662de1da921\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128550 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhkzj\" (UniqueName: \"kubernetes.io/projected/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-kube-api-access-fhkzj\") pod \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128604 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ssh-key\") pod \"748f1896-9a69-406c-bc59-0662de1da921\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128654 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-ovn-metadata-agent-neutron-config-0\") pod \"748f1896-9a69-406c-bc59-0662de1da921\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128682 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-inventory\") pod \"748f1896-9a69-406c-bc59-0662de1da921\" (UID: \"748f1896-9a69-406c-bc59-0662de1da921\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128719 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-utilities\") pod \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.128798 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-catalog-content\") pod \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\" (UID: \"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0\") " Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.133497 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-utilities" (OuterVolumeSpecName: "utilities") pod "6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" (UID: "6afe7d5c-5528-4f99-97e7-8fd0c519dfb0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.136654 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "748f1896-9a69-406c-bc59-0662de1da921" (UID: "748f1896-9a69-406c-bc59-0662de1da921"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.137046 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-kube-api-access-fhkzj" (OuterVolumeSpecName: "kube-api-access-fhkzj") pod "6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" (UID: "6afe7d5c-5528-4f99-97e7-8fd0c519dfb0"). InnerVolumeSpecName "kube-api-access-fhkzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.139618 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/748f1896-9a69-406c-bc59-0662de1da921-kube-api-access-blq4n" (OuterVolumeSpecName: "kube-api-access-blq4n") pod "748f1896-9a69-406c-bc59-0662de1da921" (UID: "748f1896-9a69-406c-bc59-0662de1da921"). InnerVolumeSpecName "kube-api-access-blq4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.146777 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ceph" (OuterVolumeSpecName: "ceph") pod "748f1896-9a69-406c-bc59-0662de1da921" (UID: "748f1896-9a69-406c-bc59-0662de1da921"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.187215 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" (UID: "6afe7d5c-5528-4f99-97e7-8fd0c519dfb0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.188497 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-inventory" (OuterVolumeSpecName: "inventory") pod "748f1896-9a69-406c-bc59-0662de1da921" (UID: "748f1896-9a69-406c-bc59-0662de1da921"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.197302 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "748f1896-9a69-406c-bc59-0662de1da921" (UID: "748f1896-9a69-406c-bc59-0662de1da921"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.201114 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "748f1896-9a69-406c-bc59-0662de1da921" (UID: "748f1896-9a69-406c-bc59-0662de1da921"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.220402 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "748f1896-9a69-406c-bc59-0662de1da921" (UID: "748f1896-9a69-406c-bc59-0662de1da921"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231003 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blq4n\" (UniqueName: \"kubernetes.io/projected/748f1896-9a69-406c-bc59-0662de1da921-kube-api-access-blq4n\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231047 4912 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231059 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231070 4912 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231083 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhkzj\" (UniqueName: \"kubernetes.io/projected/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-kube-api-access-fhkzj\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231097 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231108 4912 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231124 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/748f1896-9a69-406c-bc59-0662de1da921-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231136 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.231147 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.492180 4912 generic.go:334] "Generic (PLEG): container finished" podID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerID="71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f" exitCode=0 Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.492263 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tw66" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.492280 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tw66" event={"ID":"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0","Type":"ContainerDied","Data":"71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f"} Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.492393 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tw66" event={"ID":"6afe7d5c-5528-4f99-97e7-8fd0c519dfb0","Type":"ContainerDied","Data":"eeaa9a1bb1f7e843e5d247228522154f98a6a532a247f2bf2e914de11ea018cb"} Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.492471 4912 scope.go:117] "RemoveContainer" containerID="71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.495337 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" event={"ID":"748f1896-9a69-406c-bc59-0662de1da921","Type":"ContainerDied","Data":"e97d456edb3e067faeece2f9cefc0a2ebe4b0c2f7167c1d4890accbed233a737"} Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.495364 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e97d456edb3e067faeece2f9cefc0a2ebe4b0c2f7167c1d4890accbed233a737" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.495426 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.533291 4912 scope.go:117] "RemoveContainer" containerID="a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.579855 4912 scope.go:117] "RemoveContainer" containerID="ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.598584 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tw66"] Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.619474 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tw66"] Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.657621 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl"] Dec 03 01:33:01 crc kubenswrapper[4912]: E1203 01:33:01.658853 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="748f1896-9a69-406c-bc59-0662de1da921" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.658880 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="748f1896-9a69-406c-bc59-0662de1da921" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 01:33:01 crc kubenswrapper[4912]: E1203 01:33:01.658935 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerName="registry-server" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.658944 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerName="registry-server" Dec 03 01:33:01 crc kubenswrapper[4912]: E1203 01:33:01.658979 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerName="extract-content" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.658992 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerName="extract-content" Dec 03 01:33:01 crc kubenswrapper[4912]: E1203 01:33:01.659063 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerName="extract-utilities" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.659075 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerName="extract-utilities" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.659795 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" containerName="registry-server" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.659837 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="748f1896-9a69-406c-bc59-0662de1da921" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.661267 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.665174 4912 scope.go:117] "RemoveContainer" containerID="71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f" Dec 03 01:33:01 crc kubenswrapper[4912]: E1203 01:33:01.665686 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f\": container with ID starting with 71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f not found: ID does not exist" containerID="71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.665738 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f"} err="failed to get container status \"71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f\": rpc error: code = NotFound desc = could not find container \"71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f\": container with ID starting with 71c5d90c458e61ab0b63fbfb1718c03148e9700a420a60b50b10f3d2d24dce6f not found: ID does not exist" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.665779 4912 scope.go:117] "RemoveContainer" containerID="a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e" Dec 03 01:33:01 crc kubenswrapper[4912]: E1203 01:33:01.666607 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e\": container with ID starting with a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e not found: ID does not exist" containerID="a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.666881 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e"} err="failed to get container status \"a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e\": rpc error: code = NotFound desc = could not find container \"a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e\": container with ID starting with a05ffd6cf8c359b6081c9382c95f7abddb5e43341a2be135f39a1c05fbe1a68e not found: ID does not exist" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.666917 4912 scope.go:117] "RemoveContainer" containerID="ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446" Dec 03 01:33:01 crc kubenswrapper[4912]: E1203 01:33:01.667309 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446\": container with ID starting with ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446 not found: ID does not exist" containerID="ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.667370 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446"} err="failed to get container status \"ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446\": rpc error: code = NotFound desc = could not find container \"ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446\": container with ID starting with ebb157ef9e997d1962ad4e495dbaae2ec4fca60ed9df40cce75dc390b35ec446 not found: ID does not exist" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.667939 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.668169 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.668176 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.668310 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.668343 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.668473 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.687613 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl"] Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.755057 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.755101 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.755155 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.755330 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wztk2\" (UniqueName: \"kubernetes.io/projected/911b9678-7127-4ae0-ba4b-c059dc13796f-kube-api-access-wztk2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.755396 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.755640 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.857496 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.857789 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.857883 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.857944 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.858031 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.858096 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wztk2\" (UniqueName: \"kubernetes.io/projected/911b9678-7127-4ae0-ba4b-c059dc13796f-kube-api-access-wztk2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.861732 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.863547 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.865126 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.866135 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.866862 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:01 crc kubenswrapper[4912]: I1203 01:33:01.873629 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wztk2\" (UniqueName: \"kubernetes.io/projected/911b9678-7127-4ae0-ba4b-c059dc13796f-kube-api-access-wztk2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-q9btl\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:02 crc kubenswrapper[4912]: I1203 01:33:02.007520 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:33:02 crc kubenswrapper[4912]: W1203 01:33:02.597962 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod911b9678_7127_4ae0_ba4b_c059dc13796f.slice/crio-0eb46b417365e8b636634330664bd7ead5cd3612173018e68ca62539ebb498dc WatchSource:0}: Error finding container 0eb46b417365e8b636634330664bd7ead5cd3612173018e68ca62539ebb498dc: Status 404 returned error can't find the container with id 0eb46b417365e8b636634330664bd7ead5cd3612173018e68ca62539ebb498dc Dec 03 01:33:02 crc kubenswrapper[4912]: I1203 01:33:02.604276 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6afe7d5c-5528-4f99-97e7-8fd0c519dfb0" path="/var/lib/kubelet/pods/6afe7d5c-5528-4f99-97e7-8fd0c519dfb0/volumes" Dec 03 01:33:02 crc kubenswrapper[4912]: I1203 01:33:02.606770 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl"] Dec 03 01:33:03 crc kubenswrapper[4912]: I1203 01:33:03.532112 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" event={"ID":"911b9678-7127-4ae0-ba4b-c059dc13796f","Type":"ContainerStarted","Data":"22c103f677f4979ada048666c555caa003429160a1c27a61ad4e6243ba0e5e1e"} Dec 03 01:33:03 crc kubenswrapper[4912]: I1203 01:33:03.532617 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" event={"ID":"911b9678-7127-4ae0-ba4b-c059dc13796f","Type":"ContainerStarted","Data":"0eb46b417365e8b636634330664bd7ead5cd3612173018e68ca62539ebb498dc"} Dec 03 01:33:03 crc kubenswrapper[4912]: I1203 01:33:03.564375 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" podStartSLOduration=2.017396138 podStartE2EDuration="2.564352126s" podCreationTimestamp="2025-12-03 01:33:01 +0000 UTC" firstStartedPulling="2025-12-03 01:33:02.608317271 +0000 UTC m=+4168.250337831" lastFinishedPulling="2025-12-03 01:33:03.155273249 +0000 UTC m=+4168.797293819" observedRunningTime="2025-12-03 01:33:03.561400261 +0000 UTC m=+4169.203420851" watchObservedRunningTime="2025-12-03 01:33:03.564352126 +0000 UTC m=+4169.206372696" Dec 03 01:33:09 crc kubenswrapper[4912]: I1203 01:33:09.913604 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jb97m"] Dec 03 01:33:09 crc kubenswrapper[4912]: I1203 01:33:09.919193 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:09 crc kubenswrapper[4912]: I1203 01:33:09.932294 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jb97m"] Dec 03 01:33:09 crc kubenswrapper[4912]: I1203 01:33:09.960286 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69mz5\" (UniqueName: \"kubernetes.io/projected/d4055d3b-f9fb-430a-8471-5478bae34a1c-kube-api-access-69mz5\") pod \"community-operators-jb97m\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:09 crc kubenswrapper[4912]: I1203 01:33:09.960356 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-catalog-content\") pod \"community-operators-jb97m\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:09 crc kubenswrapper[4912]: I1203 01:33:09.960547 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-utilities\") pod \"community-operators-jb97m\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:10 crc kubenswrapper[4912]: I1203 01:33:10.061840 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-utilities\") pod \"community-operators-jb97m\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:10 crc kubenswrapper[4912]: I1203 01:33:10.061957 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69mz5\" (UniqueName: \"kubernetes.io/projected/d4055d3b-f9fb-430a-8471-5478bae34a1c-kube-api-access-69mz5\") pod \"community-operators-jb97m\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:10 crc kubenswrapper[4912]: I1203 01:33:10.061995 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-catalog-content\") pod \"community-operators-jb97m\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:10 crc kubenswrapper[4912]: I1203 01:33:10.062494 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-utilities\") pod \"community-operators-jb97m\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:10 crc kubenswrapper[4912]: I1203 01:33:10.062547 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-catalog-content\") pod \"community-operators-jb97m\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:10 crc kubenswrapper[4912]: I1203 01:33:10.084589 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69mz5\" (UniqueName: \"kubernetes.io/projected/d4055d3b-f9fb-430a-8471-5478bae34a1c-kube-api-access-69mz5\") pod \"community-operators-jb97m\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:10 crc kubenswrapper[4912]: I1203 01:33:10.256148 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:11 crc kubenswrapper[4912]: I1203 01:33:11.275297 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jb97m"] Dec 03 01:33:11 crc kubenswrapper[4912]: I1203 01:33:11.631411 4912 generic.go:334] "Generic (PLEG): container finished" podID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerID="1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220" exitCode=0 Dec 03 01:33:11 crc kubenswrapper[4912]: I1203 01:33:11.632081 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jb97m" event={"ID":"d4055d3b-f9fb-430a-8471-5478bae34a1c","Type":"ContainerDied","Data":"1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220"} Dec 03 01:33:11 crc kubenswrapper[4912]: I1203 01:33:11.632133 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jb97m" event={"ID":"d4055d3b-f9fb-430a-8471-5478bae34a1c","Type":"ContainerStarted","Data":"be5e221859e7a8312e9e29e776fc68bb3e3415b8a443e0682fcfc93cd4714956"} Dec 03 01:33:13 crc kubenswrapper[4912]: I1203 01:33:13.660216 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jb97m" event={"ID":"d4055d3b-f9fb-430a-8471-5478bae34a1c","Type":"ContainerStarted","Data":"0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b"} Dec 03 01:33:14 crc kubenswrapper[4912]: I1203 01:33:14.674661 4912 generic.go:334] "Generic (PLEG): container finished" podID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerID="0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b" exitCode=0 Dec 03 01:33:14 crc kubenswrapper[4912]: I1203 01:33:14.674705 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jb97m" event={"ID":"d4055d3b-f9fb-430a-8471-5478bae34a1c","Type":"ContainerDied","Data":"0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b"} Dec 03 01:33:15 crc kubenswrapper[4912]: I1203 01:33:15.690764 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jb97m" event={"ID":"d4055d3b-f9fb-430a-8471-5478bae34a1c","Type":"ContainerStarted","Data":"61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f"} Dec 03 01:33:15 crc kubenswrapper[4912]: I1203 01:33:15.718170 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jb97m" podStartSLOduration=3.236440612 podStartE2EDuration="6.718145122s" podCreationTimestamp="2025-12-03 01:33:09 +0000 UTC" firstStartedPulling="2025-12-03 01:33:11.635579965 +0000 UTC m=+4177.277600535" lastFinishedPulling="2025-12-03 01:33:15.117284485 +0000 UTC m=+4180.759305045" observedRunningTime="2025-12-03 01:33:15.710148771 +0000 UTC m=+4181.352169331" watchObservedRunningTime="2025-12-03 01:33:15.718145122 +0000 UTC m=+4181.360165682" Dec 03 01:33:20 crc kubenswrapper[4912]: I1203 01:33:20.257083 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:20 crc kubenswrapper[4912]: I1203 01:33:20.257612 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:20 crc kubenswrapper[4912]: I1203 01:33:20.713831 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:20 crc kubenswrapper[4912]: I1203 01:33:20.803825 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:20 crc kubenswrapper[4912]: I1203 01:33:20.960677 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jb97m"] Dec 03 01:33:22 crc kubenswrapper[4912]: I1203 01:33:22.776417 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jb97m" podUID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerName="registry-server" containerID="cri-o://61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f" gracePeriod=2 Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.340277 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.465279 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-catalog-content\") pod \"d4055d3b-f9fb-430a-8471-5478bae34a1c\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.465767 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-utilities\") pod \"d4055d3b-f9fb-430a-8471-5478bae34a1c\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.466161 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69mz5\" (UniqueName: \"kubernetes.io/projected/d4055d3b-f9fb-430a-8471-5478bae34a1c-kube-api-access-69mz5\") pod \"d4055d3b-f9fb-430a-8471-5478bae34a1c\" (UID: \"d4055d3b-f9fb-430a-8471-5478bae34a1c\") " Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.466645 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-utilities" (OuterVolumeSpecName: "utilities") pod "d4055d3b-f9fb-430a-8471-5478bae34a1c" (UID: "d4055d3b-f9fb-430a-8471-5478bae34a1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.471928 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4055d3b-f9fb-430a-8471-5478bae34a1c-kube-api-access-69mz5" (OuterVolumeSpecName: "kube-api-access-69mz5") pod "d4055d3b-f9fb-430a-8471-5478bae34a1c" (UID: "d4055d3b-f9fb-430a-8471-5478bae34a1c"). InnerVolumeSpecName "kube-api-access-69mz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.545340 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4055d3b-f9fb-430a-8471-5478bae34a1c" (UID: "d4055d3b-f9fb-430a-8471-5478bae34a1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.568597 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69mz5\" (UniqueName: \"kubernetes.io/projected/d4055d3b-f9fb-430a-8471-5478bae34a1c-kube-api-access-69mz5\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.568635 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.568646 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4055d3b-f9fb-430a-8471-5478bae34a1c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.794943 4912 generic.go:334] "Generic (PLEG): container finished" podID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerID="61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f" exitCode=0 Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.795004 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jb97m" event={"ID":"d4055d3b-f9fb-430a-8471-5478bae34a1c","Type":"ContainerDied","Data":"61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f"} Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.795046 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jb97m" event={"ID":"d4055d3b-f9fb-430a-8471-5478bae34a1c","Type":"ContainerDied","Data":"be5e221859e7a8312e9e29e776fc68bb3e3415b8a443e0682fcfc93cd4714956"} Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.795074 4912 scope.go:117] "RemoveContainer" containerID="61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.795266 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jb97m" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.830077 4912 scope.go:117] "RemoveContainer" containerID="0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.851763 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jb97m"] Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.864480 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jb97m"] Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.865946 4912 scope.go:117] "RemoveContainer" containerID="1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.914127 4912 scope.go:117] "RemoveContainer" containerID="61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f" Dec 03 01:33:23 crc kubenswrapper[4912]: E1203 01:33:23.914726 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f\": container with ID starting with 61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f not found: ID does not exist" containerID="61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.914889 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f"} err="failed to get container status \"61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f\": rpc error: code = NotFound desc = could not find container \"61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f\": container with ID starting with 61553b076f0017e973e09179b743bac304ec9b94e6539b397dc91706f2f5ec1f not found: ID does not exist" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.915006 4912 scope.go:117] "RemoveContainer" containerID="0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b" Dec 03 01:33:23 crc kubenswrapper[4912]: E1203 01:33:23.915527 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b\": container with ID starting with 0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b not found: ID does not exist" containerID="0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.915699 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b"} err="failed to get container status \"0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b\": rpc error: code = NotFound desc = could not find container \"0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b\": container with ID starting with 0a60c6fe39970f636218f6fbcd7a306e16bb54c0932bca849d50e21c7c8aa66b not found: ID does not exist" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.915834 4912 scope.go:117] "RemoveContainer" containerID="1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220" Dec 03 01:33:23 crc kubenswrapper[4912]: E1203 01:33:23.916197 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220\": container with ID starting with 1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220 not found: ID does not exist" containerID="1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220" Dec 03 01:33:23 crc kubenswrapper[4912]: I1203 01:33:23.916317 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220"} err="failed to get container status \"1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220\": rpc error: code = NotFound desc = could not find container \"1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220\": container with ID starting with 1d29f88d29e2b2865d8e6ae7d3cb49c05040181c53f9db9e934288353d82b220 not found: ID does not exist" Dec 03 01:33:24 crc kubenswrapper[4912]: I1203 01:33:24.592260 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4055d3b-f9fb-430a-8471-5478bae34a1c" path="/var/lib/kubelet/pods/d4055d3b-f9fb-430a-8471-5478bae34a1c/volumes" Dec 03 01:33:25 crc kubenswrapper[4912]: I1203 01:33:25.467687 4912 scope.go:117] "RemoveContainer" containerID="5838146184a31e7553db7f0db9004633ce0fc4b73923a1829c23ea8b725bf305" Dec 03 01:33:25 crc kubenswrapper[4912]: I1203 01:33:25.509579 4912 scope.go:117] "RemoveContainer" containerID="b8ea703087999ce7fba46385652f0a5e751ce30f21a6cf4ff921fbeb349c18bf" Dec 03 01:33:25 crc kubenswrapper[4912]: I1203 01:33:25.569819 4912 scope.go:117] "RemoveContainer" containerID="27098b1be5e072fb28fc23e9974687aefb3c124888a993c0394f47228d346e75" Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.881157 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ndfph"] Dec 03 01:33:51 crc kubenswrapper[4912]: E1203 01:33:51.882523 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerName="extract-utilities" Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.882547 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerName="extract-utilities" Dec 03 01:33:51 crc kubenswrapper[4912]: E1203 01:33:51.882591 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerName="extract-content" Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.882602 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerName="extract-content" Dec 03 01:33:51 crc kubenswrapper[4912]: E1203 01:33:51.882626 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerName="registry-server" Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.882636 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerName="registry-server" Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.882994 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4055d3b-f9fb-430a-8471-5478bae34a1c" containerName="registry-server" Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.885681 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.929794 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ndfph"] Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.944958 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-utilities\") pod \"redhat-operators-ndfph\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.948953 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cd79\" (UniqueName: \"kubernetes.io/projected/480ca3e0-d6fe-4e95-a217-84173a149172-kube-api-access-8cd79\") pod \"redhat-operators-ndfph\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:51 crc kubenswrapper[4912]: I1203 01:33:51.949096 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-catalog-content\") pod \"redhat-operators-ndfph\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:52 crc kubenswrapper[4912]: I1203 01:33:52.052174 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-utilities\") pod \"redhat-operators-ndfph\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:52 crc kubenswrapper[4912]: I1203 01:33:52.052282 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cd79\" (UniqueName: \"kubernetes.io/projected/480ca3e0-d6fe-4e95-a217-84173a149172-kube-api-access-8cd79\") pod \"redhat-operators-ndfph\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:52 crc kubenswrapper[4912]: I1203 01:33:52.052325 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-catalog-content\") pod \"redhat-operators-ndfph\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:52 crc kubenswrapper[4912]: I1203 01:33:52.052951 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-catalog-content\") pod \"redhat-operators-ndfph\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:52 crc kubenswrapper[4912]: I1203 01:33:52.052942 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-utilities\") pod \"redhat-operators-ndfph\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:52 crc kubenswrapper[4912]: I1203 01:33:52.076000 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cd79\" (UniqueName: \"kubernetes.io/projected/480ca3e0-d6fe-4e95-a217-84173a149172-kube-api-access-8cd79\") pod \"redhat-operators-ndfph\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:52 crc kubenswrapper[4912]: I1203 01:33:52.228297 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:33:52 crc kubenswrapper[4912]: I1203 01:33:52.717853 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ndfph"] Dec 03 01:33:53 crc kubenswrapper[4912]: I1203 01:33:53.329082 4912 generic.go:334] "Generic (PLEG): container finished" podID="480ca3e0-d6fe-4e95-a217-84173a149172" containerID="97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c" exitCode=0 Dec 03 01:33:53 crc kubenswrapper[4912]: I1203 01:33:53.329196 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndfph" event={"ID":"480ca3e0-d6fe-4e95-a217-84173a149172","Type":"ContainerDied","Data":"97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c"} Dec 03 01:33:53 crc kubenswrapper[4912]: I1203 01:33:53.329619 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndfph" event={"ID":"480ca3e0-d6fe-4e95-a217-84173a149172","Type":"ContainerStarted","Data":"e8fb0ae37fecd4bf31e7ef1dceab2e1bf5a378db80a303b442f92d002a570b2c"} Dec 03 01:33:54 crc kubenswrapper[4912]: I1203 01:33:54.348504 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndfph" event={"ID":"480ca3e0-d6fe-4e95-a217-84173a149172","Type":"ContainerStarted","Data":"411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb"} Dec 03 01:33:56 crc kubenswrapper[4912]: I1203 01:33:56.374552 4912 generic.go:334] "Generic (PLEG): container finished" podID="480ca3e0-d6fe-4e95-a217-84173a149172" containerID="411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb" exitCode=0 Dec 03 01:33:56 crc kubenswrapper[4912]: I1203 01:33:56.374621 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndfph" event={"ID":"480ca3e0-d6fe-4e95-a217-84173a149172","Type":"ContainerDied","Data":"411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb"} Dec 03 01:33:57 crc kubenswrapper[4912]: I1203 01:33:57.399936 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndfph" event={"ID":"480ca3e0-d6fe-4e95-a217-84173a149172","Type":"ContainerStarted","Data":"dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb"} Dec 03 01:33:57 crc kubenswrapper[4912]: I1203 01:33:57.422378 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ndfph" podStartSLOduration=2.702290571 podStartE2EDuration="6.422363969s" podCreationTimestamp="2025-12-03 01:33:51 +0000 UTC" firstStartedPulling="2025-12-03 01:33:53.331383568 +0000 UTC m=+4218.973404128" lastFinishedPulling="2025-12-03 01:33:57.051456946 +0000 UTC m=+4222.693477526" observedRunningTime="2025-12-03 01:33:57.415903033 +0000 UTC m=+4223.057923593" watchObservedRunningTime="2025-12-03 01:33:57.422363969 +0000 UTC m=+4223.064384529" Dec 03 01:34:02 crc kubenswrapper[4912]: I1203 01:34:02.228520 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:34:02 crc kubenswrapper[4912]: I1203 01:34:02.229710 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:34:03 crc kubenswrapper[4912]: I1203 01:34:03.293601 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ndfph" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" containerName="registry-server" probeResult="failure" output=< Dec 03 01:34:03 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 01:34:03 crc kubenswrapper[4912]: > Dec 03 01:34:12 crc kubenswrapper[4912]: I1203 01:34:12.325776 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:34:12 crc kubenswrapper[4912]: I1203 01:34:12.406570 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:34:12 crc kubenswrapper[4912]: I1203 01:34:12.595967 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ndfph"] Dec 03 01:34:13 crc kubenswrapper[4912]: I1203 01:34:13.620534 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ndfph" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" containerName="registry-server" containerID="cri-o://dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb" gracePeriod=2 Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.144869 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.249219 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-utilities\") pod \"480ca3e0-d6fe-4e95-a217-84173a149172\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.249285 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-catalog-content\") pod \"480ca3e0-d6fe-4e95-a217-84173a149172\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.249337 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cd79\" (UniqueName: \"kubernetes.io/projected/480ca3e0-d6fe-4e95-a217-84173a149172-kube-api-access-8cd79\") pod \"480ca3e0-d6fe-4e95-a217-84173a149172\" (UID: \"480ca3e0-d6fe-4e95-a217-84173a149172\") " Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.256505 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-utilities" (OuterVolumeSpecName: "utilities") pod "480ca3e0-d6fe-4e95-a217-84173a149172" (UID: "480ca3e0-d6fe-4e95-a217-84173a149172"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.258558 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480ca3e0-d6fe-4e95-a217-84173a149172-kube-api-access-8cd79" (OuterVolumeSpecName: "kube-api-access-8cd79") pod "480ca3e0-d6fe-4e95-a217-84173a149172" (UID: "480ca3e0-d6fe-4e95-a217-84173a149172"). InnerVolumeSpecName "kube-api-access-8cd79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.352683 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.352726 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cd79\" (UniqueName: \"kubernetes.io/projected/480ca3e0-d6fe-4e95-a217-84173a149172-kube-api-access-8cd79\") on node \"crc\" DevicePath \"\"" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.358022 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "480ca3e0-d6fe-4e95-a217-84173a149172" (UID: "480ca3e0-d6fe-4e95-a217-84173a149172"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.455324 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ca3e0-d6fe-4e95-a217-84173a149172-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.637294 4912 generic.go:334] "Generic (PLEG): container finished" podID="480ca3e0-d6fe-4e95-a217-84173a149172" containerID="dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb" exitCode=0 Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.637354 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndfph" event={"ID":"480ca3e0-d6fe-4e95-a217-84173a149172","Type":"ContainerDied","Data":"dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb"} Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.637397 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ndfph" event={"ID":"480ca3e0-d6fe-4e95-a217-84173a149172","Type":"ContainerDied","Data":"e8fb0ae37fecd4bf31e7ef1dceab2e1bf5a378db80a303b442f92d002a570b2c"} Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.637426 4912 scope.go:117] "RemoveContainer" containerID="dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.637639 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ndfph" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.686311 4912 scope.go:117] "RemoveContainer" containerID="411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.688277 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ndfph"] Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.709265 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ndfph"] Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.724283 4912 scope.go:117] "RemoveContainer" containerID="97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.778324 4912 scope.go:117] "RemoveContainer" containerID="dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb" Dec 03 01:34:14 crc kubenswrapper[4912]: E1203 01:34:14.778859 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb\": container with ID starting with dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb not found: ID does not exist" containerID="dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.778894 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb"} err="failed to get container status \"dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb\": rpc error: code = NotFound desc = could not find container \"dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb\": container with ID starting with dea1d68554da677682518a65ad05bf970999dfb5db042266ae510e078dab56fb not found: ID does not exist" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.778920 4912 scope.go:117] "RemoveContainer" containerID="411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb" Dec 03 01:34:14 crc kubenswrapper[4912]: E1203 01:34:14.779272 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb\": container with ID starting with 411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb not found: ID does not exist" containerID="411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.779329 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb"} err="failed to get container status \"411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb\": rpc error: code = NotFound desc = could not find container \"411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb\": container with ID starting with 411f354d2a0e1105c986c29d0e68e7a2897cb1d8ca6e0591ed5f26d026a205cb not found: ID does not exist" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.779365 4912 scope.go:117] "RemoveContainer" containerID="97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c" Dec 03 01:34:14 crc kubenswrapper[4912]: E1203 01:34:14.779721 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c\": container with ID starting with 97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c not found: ID does not exist" containerID="97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c" Dec 03 01:34:14 crc kubenswrapper[4912]: I1203 01:34:14.779753 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c"} err="failed to get container status \"97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c\": rpc error: code = NotFound desc = could not find container \"97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c\": container with ID starting with 97677238ef3c34d3232e0441deeaf1c6fca8d4f4ddbc7eac2b526d650a61b93c not found: ID does not exist" Dec 03 01:34:16 crc kubenswrapper[4912]: I1203 01:34:16.589254 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" path="/var/lib/kubelet/pods/480ca3e0-d6fe-4e95-a217-84173a149172/volumes" Dec 03 01:34:48 crc kubenswrapper[4912]: I1203 01:34:48.077861 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:34:48 crc kubenswrapper[4912]: I1203 01:34:48.078387 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:35:18 crc kubenswrapper[4912]: I1203 01:35:18.078701 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:35:18 crc kubenswrapper[4912]: I1203 01:35:18.079367 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:35:48 crc kubenswrapper[4912]: I1203 01:35:48.077616 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:35:48 crc kubenswrapper[4912]: I1203 01:35:48.078272 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:35:48 crc kubenswrapper[4912]: I1203 01:35:48.078335 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:35:48 crc kubenswrapper[4912]: I1203 01:35:48.079577 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:35:48 crc kubenswrapper[4912]: I1203 01:35:48.079674 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" gracePeriod=600 Dec 03 01:35:48 crc kubenswrapper[4912]: E1203 01:35:48.209341 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:35:48 crc kubenswrapper[4912]: I1203 01:35:48.777386 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" exitCode=0 Dec 03 01:35:48 crc kubenswrapper[4912]: I1203 01:35:48.777509 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00"} Dec 03 01:35:48 crc kubenswrapper[4912]: I1203 01:35:48.778126 4912 scope.go:117] "RemoveContainer" containerID="f5ddc05f695c6bfb74ae5cda3f4bce456d3a1a95d24096442007066627b0d1c3" Dec 03 01:35:48 crc kubenswrapper[4912]: I1203 01:35:48.779653 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:35:48 crc kubenswrapper[4912]: E1203 01:35:48.780623 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:36:02 crc kubenswrapper[4912]: I1203 01:36:02.574415 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:36:02 crc kubenswrapper[4912]: E1203 01:36:02.576066 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:36:16 crc kubenswrapper[4912]: I1203 01:36:16.572421 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:36:16 crc kubenswrapper[4912]: E1203 01:36:16.573271 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:36:30 crc kubenswrapper[4912]: I1203 01:36:30.572978 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:36:30 crc kubenswrapper[4912]: E1203 01:36:30.573913 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:36:45 crc kubenswrapper[4912]: I1203 01:36:45.573181 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:36:45 crc kubenswrapper[4912]: E1203 01:36:45.574060 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:36:55 crc kubenswrapper[4912]: I1203 01:36:55.548724 4912 generic.go:334] "Generic (PLEG): container finished" podID="911b9678-7127-4ae0-ba4b-c059dc13796f" containerID="22c103f677f4979ada048666c555caa003429160a1c27a61ad4e6243ba0e5e1e" exitCode=0 Dec 03 01:36:55 crc kubenswrapper[4912]: I1203 01:36:55.548916 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" event={"ID":"911b9678-7127-4ae0-ba4b-c059dc13796f","Type":"ContainerDied","Data":"22c103f677f4979ada048666c555caa003429160a1c27a61ad4e6243ba0e5e1e"} Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.165220 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.267394 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wztk2\" (UniqueName: \"kubernetes.io/projected/911b9678-7127-4ae0-ba4b-c059dc13796f-kube-api-access-wztk2\") pod \"911b9678-7127-4ae0-ba4b-c059dc13796f\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.267491 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ssh-key\") pod \"911b9678-7127-4ae0-ba4b-c059dc13796f\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.267602 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-inventory\") pod \"911b9678-7127-4ae0-ba4b-c059dc13796f\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.267634 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-combined-ca-bundle\") pod \"911b9678-7127-4ae0-ba4b-c059dc13796f\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.267729 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-secret-0\") pod \"911b9678-7127-4ae0-ba4b-c059dc13796f\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.267782 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ceph\") pod \"911b9678-7127-4ae0-ba4b-c059dc13796f\" (UID: \"911b9678-7127-4ae0-ba4b-c059dc13796f\") " Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.282307 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ceph" (OuterVolumeSpecName: "ceph") pod "911b9678-7127-4ae0-ba4b-c059dc13796f" (UID: "911b9678-7127-4ae0-ba4b-c059dc13796f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.284378 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/911b9678-7127-4ae0-ba4b-c059dc13796f-kube-api-access-wztk2" (OuterVolumeSpecName: "kube-api-access-wztk2") pod "911b9678-7127-4ae0-ba4b-c059dc13796f" (UID: "911b9678-7127-4ae0-ba4b-c059dc13796f"). InnerVolumeSpecName "kube-api-access-wztk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.297999 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "911b9678-7127-4ae0-ba4b-c059dc13796f" (UID: "911b9678-7127-4ae0-ba4b-c059dc13796f"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.314677 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "911b9678-7127-4ae0-ba4b-c059dc13796f" (UID: "911b9678-7127-4ae0-ba4b-c059dc13796f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.316596 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-inventory" (OuterVolumeSpecName: "inventory") pod "911b9678-7127-4ae0-ba4b-c059dc13796f" (UID: "911b9678-7127-4ae0-ba4b-c059dc13796f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.316799 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "911b9678-7127-4ae0-ba4b-c059dc13796f" (UID: "911b9678-7127-4ae0-ba4b-c059dc13796f"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.371273 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wztk2\" (UniqueName: \"kubernetes.io/projected/911b9678-7127-4ae0-ba4b-c059dc13796f-kube-api-access-wztk2\") on node \"crc\" DevicePath \"\"" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.371319 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.371330 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.371342 4912 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.371354 4912 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.371366 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/911b9678-7127-4ae0-ba4b-c059dc13796f-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.580793 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" event={"ID":"911b9678-7127-4ae0-ba4b-c059dc13796f","Type":"ContainerDied","Data":"0eb46b417365e8b636634330664bd7ead5cd3612173018e68ca62539ebb498dc"} Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.580845 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0eb46b417365e8b636634330664bd7ead5cd3612173018e68ca62539ebb498dc" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.580922 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-q9btl" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.741726 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8"] Dec 03 01:36:57 crc kubenswrapper[4912]: E1203 01:36:57.742209 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" containerName="registry-server" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.742229 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" containerName="registry-server" Dec 03 01:36:57 crc kubenswrapper[4912]: E1203 01:36:57.742251 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" containerName="extract-content" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.742260 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" containerName="extract-content" Dec 03 01:36:57 crc kubenswrapper[4912]: E1203 01:36:57.742279 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" containerName="extract-utilities" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.742287 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" containerName="extract-utilities" Dec 03 01:36:57 crc kubenswrapper[4912]: E1203 01:36:57.742308 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="911b9678-7127-4ae0-ba4b-c059dc13796f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.742318 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="911b9678-7127-4ae0-ba4b-c059dc13796f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.742584 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="911b9678-7127-4ae0-ba4b-c059dc13796f" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.742612 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="480ca3e0-d6fe-4e95-a217-84173a149172" containerName="registry-server" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.744467 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.746604 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.748160 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.748360 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.748727 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.748868 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.748978 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.750006 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.750253 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.752781 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.754296 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8"] Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.879183 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.879242 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj8pb\" (UniqueName: \"kubernetes.io/projected/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-kube-api-access-jj8pb\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.879271 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.879454 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.879604 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.879798 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.879961 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.880098 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.880296 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.880366 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.880590 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.982522 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.982616 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj8pb\" (UniqueName: \"kubernetes.io/projected/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-kube-api-access-jj8pb\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.982669 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.982712 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.982781 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.982816 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.982860 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.982899 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.982978 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.983013 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.983103 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.985230 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.988104 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.988837 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.989000 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.989248 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.989576 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.991172 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.992901 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.992960 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:57 crc kubenswrapper[4912]: I1203 01:36:57.993985 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:58 crc kubenswrapper[4912]: I1203 01:36:58.007998 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj8pb\" (UniqueName: \"kubernetes.io/projected/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-kube-api-access-jj8pb\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:58 crc kubenswrapper[4912]: I1203 01:36:58.075480 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:36:58 crc kubenswrapper[4912]: I1203 01:36:58.640196 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8"] Dec 03 01:36:59 crc kubenswrapper[4912]: I1203 01:36:59.602982 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" event={"ID":"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e","Type":"ContainerStarted","Data":"0b9b1ed67406363c87d79c71afe55092a01400ad840bddbda9a7d26d18cfd111"} Dec 03 01:36:59 crc kubenswrapper[4912]: I1203 01:36:59.603293 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" event={"ID":"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e","Type":"ContainerStarted","Data":"71717653c7c0332bc90e2251e4596de2151d28db95e30c9f332306afb898ca3c"} Dec 03 01:36:59 crc kubenswrapper[4912]: I1203 01:36:59.637525 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" podStartSLOduration=2.134483948 podStartE2EDuration="2.637505524s" podCreationTimestamp="2025-12-03 01:36:57 +0000 UTC" firstStartedPulling="2025-12-03 01:36:58.644919595 +0000 UTC m=+4404.286940155" lastFinishedPulling="2025-12-03 01:36:59.147941131 +0000 UTC m=+4404.789961731" observedRunningTime="2025-12-03 01:36:59.63099904 +0000 UTC m=+4405.273019620" watchObservedRunningTime="2025-12-03 01:36:59.637505524 +0000 UTC m=+4405.279526084" Dec 03 01:37:00 crc kubenswrapper[4912]: I1203 01:37:00.572553 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:37:00 crc kubenswrapper[4912]: E1203 01:37:00.573248 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:37:13 crc kubenswrapper[4912]: I1203 01:37:13.571642 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:37:13 crc kubenswrapper[4912]: E1203 01:37:13.572618 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:37:25 crc kubenswrapper[4912]: I1203 01:37:25.572914 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:37:25 crc kubenswrapper[4912]: E1203 01:37:25.573750 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:37:38 crc kubenswrapper[4912]: I1203 01:37:38.574565 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:37:38 crc kubenswrapper[4912]: E1203 01:37:38.575935 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:37:50 crc kubenswrapper[4912]: I1203 01:37:50.571890 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:37:50 crc kubenswrapper[4912]: E1203 01:37:50.572591 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:38:02 crc kubenswrapper[4912]: I1203 01:38:02.572522 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:38:02 crc kubenswrapper[4912]: E1203 01:38:02.573106 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.616322 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8gcrj"] Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.620207 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.639989 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8gcrj"] Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.788105 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfszm\" (UniqueName: \"kubernetes.io/projected/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-kube-api-access-tfszm\") pod \"certified-operators-8gcrj\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.788207 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-catalog-content\") pod \"certified-operators-8gcrj\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.788317 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-utilities\") pod \"certified-operators-8gcrj\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.890103 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfszm\" (UniqueName: \"kubernetes.io/projected/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-kube-api-access-tfszm\") pod \"certified-operators-8gcrj\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.890167 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-catalog-content\") pod \"certified-operators-8gcrj\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.890244 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-utilities\") pod \"certified-operators-8gcrj\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.890874 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-utilities\") pod \"certified-operators-8gcrj\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.890959 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-catalog-content\") pod \"certified-operators-8gcrj\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.926740 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfszm\" (UniqueName: \"kubernetes.io/projected/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-kube-api-access-tfszm\") pod \"certified-operators-8gcrj\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:05 crc kubenswrapper[4912]: I1203 01:38:05.938927 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:06 crc kubenswrapper[4912]: I1203 01:38:06.447599 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8gcrj"] Dec 03 01:38:07 crc kubenswrapper[4912]: I1203 01:38:07.454612 4912 generic.go:334] "Generic (PLEG): container finished" podID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerID="0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4" exitCode=0 Dec 03 01:38:07 crc kubenswrapper[4912]: I1203 01:38:07.454692 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gcrj" event={"ID":"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb","Type":"ContainerDied","Data":"0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4"} Dec 03 01:38:07 crc kubenswrapper[4912]: I1203 01:38:07.455044 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gcrj" event={"ID":"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb","Type":"ContainerStarted","Data":"334b3d805496d7d86313ed1f8ed166093dae24f2fee58455d2be912d225492c1"} Dec 03 01:38:07 crc kubenswrapper[4912]: I1203 01:38:07.458134 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:38:08 crc kubenswrapper[4912]: I1203 01:38:08.466360 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gcrj" event={"ID":"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb","Type":"ContainerStarted","Data":"b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc"} Dec 03 01:38:09 crc kubenswrapper[4912]: I1203 01:38:09.479846 4912 generic.go:334] "Generic (PLEG): container finished" podID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerID="b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc" exitCode=0 Dec 03 01:38:09 crc kubenswrapper[4912]: I1203 01:38:09.479911 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gcrj" event={"ID":"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb","Type":"ContainerDied","Data":"b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc"} Dec 03 01:38:10 crc kubenswrapper[4912]: I1203 01:38:10.495055 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gcrj" event={"ID":"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb","Type":"ContainerStarted","Data":"d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6"} Dec 03 01:38:10 crc kubenswrapper[4912]: I1203 01:38:10.525538 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8gcrj" podStartSLOduration=2.987223225 podStartE2EDuration="5.525512666s" podCreationTimestamp="2025-12-03 01:38:05 +0000 UTC" firstStartedPulling="2025-12-03 01:38:07.457689137 +0000 UTC m=+4473.099709727" lastFinishedPulling="2025-12-03 01:38:09.995978578 +0000 UTC m=+4475.637999168" observedRunningTime="2025-12-03 01:38:10.51890936 +0000 UTC m=+4476.160929980" watchObservedRunningTime="2025-12-03 01:38:10.525512666 +0000 UTC m=+4476.167533256" Dec 03 01:38:15 crc kubenswrapper[4912]: I1203 01:38:15.939594 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:15 crc kubenswrapper[4912]: I1203 01:38:15.940604 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:16 crc kubenswrapper[4912]: I1203 01:38:16.004566 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:16 crc kubenswrapper[4912]: I1203 01:38:16.572684 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:38:16 crc kubenswrapper[4912]: E1203 01:38:16.573717 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:38:16 crc kubenswrapper[4912]: I1203 01:38:16.639418 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:16 crc kubenswrapper[4912]: I1203 01:38:16.716236 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8gcrj"] Dec 03 01:38:18 crc kubenswrapper[4912]: I1203 01:38:18.601015 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8gcrj" podUID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerName="registry-server" containerID="cri-o://d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6" gracePeriod=2 Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.224643 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.358660 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-catalog-content\") pod \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.358948 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfszm\" (UniqueName: \"kubernetes.io/projected/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-kube-api-access-tfszm\") pod \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.359117 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-utilities\") pod \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\" (UID: \"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb\") " Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.359952 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-utilities" (OuterVolumeSpecName: "utilities") pod "c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" (UID: "c9fb0b98-6f6f-432d-829e-80c2b0cf24fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.386957 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-kube-api-access-tfszm" (OuterVolumeSpecName: "kube-api-access-tfszm") pod "c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" (UID: "c9fb0b98-6f6f-432d-829e-80c2b0cf24fb"). InnerVolumeSpecName "kube-api-access-tfszm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.415741 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" (UID: "c9fb0b98-6f6f-432d-829e-80c2b0cf24fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.462442 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.462475 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.462488 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfszm\" (UniqueName: \"kubernetes.io/projected/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb-kube-api-access-tfszm\") on node \"crc\" DevicePath \"\"" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.613312 4912 generic.go:334] "Generic (PLEG): container finished" podID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerID="d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6" exitCode=0 Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.613358 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gcrj" event={"ID":"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb","Type":"ContainerDied","Data":"d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6"} Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.613727 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8gcrj" event={"ID":"c9fb0b98-6f6f-432d-829e-80c2b0cf24fb","Type":"ContainerDied","Data":"334b3d805496d7d86313ed1f8ed166093dae24f2fee58455d2be912d225492c1"} Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.613761 4912 scope.go:117] "RemoveContainer" containerID="d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.613488 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8gcrj" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.651009 4912 scope.go:117] "RemoveContainer" containerID="b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.652644 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8gcrj"] Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.662714 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8gcrj"] Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.689017 4912 scope.go:117] "RemoveContainer" containerID="0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.729034 4912 scope.go:117] "RemoveContainer" containerID="d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6" Dec 03 01:38:19 crc kubenswrapper[4912]: E1203 01:38:19.729669 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6\": container with ID starting with d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6 not found: ID does not exist" containerID="d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.729790 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6"} err="failed to get container status \"d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6\": rpc error: code = NotFound desc = could not find container \"d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6\": container with ID starting with d51483096e86642f73e9dca2ec09c35f58f7a6c0294e327b7648d9ba2f008bf6 not found: ID does not exist" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.729814 4912 scope.go:117] "RemoveContainer" containerID="b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc" Dec 03 01:38:19 crc kubenswrapper[4912]: E1203 01:38:19.730261 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc\": container with ID starting with b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc not found: ID does not exist" containerID="b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.730315 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc"} err="failed to get container status \"b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc\": rpc error: code = NotFound desc = could not find container \"b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc\": container with ID starting with b5c6d274c22456f73611734fd90019100d58808ed92f83f4dd7fc0088cdfd9dc not found: ID does not exist" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.730341 4912 scope.go:117] "RemoveContainer" containerID="0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4" Dec 03 01:38:19 crc kubenswrapper[4912]: E1203 01:38:19.730912 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4\": container with ID starting with 0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4 not found: ID does not exist" containerID="0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4" Dec 03 01:38:19 crc kubenswrapper[4912]: I1203 01:38:19.730938 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4"} err="failed to get container status \"0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4\": rpc error: code = NotFound desc = could not find container \"0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4\": container with ID starting with 0dc26c7f83592861de79b18aa0fe9ea0fd97eeac3cda413df81380e5eaa109d4 not found: ID does not exist" Dec 03 01:38:20 crc kubenswrapper[4912]: I1203 01:38:20.614230 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" path="/var/lib/kubelet/pods/c9fb0b98-6f6f-432d-829e-80c2b0cf24fb/volumes" Dec 03 01:38:30 crc kubenswrapper[4912]: I1203 01:38:30.573897 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:38:30 crc kubenswrapper[4912]: E1203 01:38:30.575087 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:38:44 crc kubenswrapper[4912]: I1203 01:38:44.585621 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:38:44 crc kubenswrapper[4912]: E1203 01:38:44.587060 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:38:57 crc kubenswrapper[4912]: I1203 01:38:57.572968 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:38:57 crc kubenswrapper[4912]: E1203 01:38:57.574099 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:39:09 crc kubenswrapper[4912]: I1203 01:39:09.573066 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:39:09 crc kubenswrapper[4912]: E1203 01:39:09.574190 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:39:22 crc kubenswrapper[4912]: I1203 01:39:22.572299 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:39:22 crc kubenswrapper[4912]: E1203 01:39:22.573345 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:39:37 crc kubenswrapper[4912]: I1203 01:39:37.573652 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:39:37 crc kubenswrapper[4912]: E1203 01:39:37.575661 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:39:51 crc kubenswrapper[4912]: I1203 01:39:51.572219 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:39:51 crc kubenswrapper[4912]: E1203 01:39:51.573007 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:40:05 crc kubenswrapper[4912]: I1203 01:40:05.571486 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:40:05 crc kubenswrapper[4912]: E1203 01:40:05.572090 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:40:19 crc kubenswrapper[4912]: I1203 01:40:19.572614 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:40:19 crc kubenswrapper[4912]: E1203 01:40:19.573666 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:40:31 crc kubenswrapper[4912]: I1203 01:40:31.572501 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:40:31 crc kubenswrapper[4912]: E1203 01:40:31.574349 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:40:45 crc kubenswrapper[4912]: I1203 01:40:45.572376 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:40:45 crc kubenswrapper[4912]: E1203 01:40:45.573016 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:40:59 crc kubenswrapper[4912]: I1203 01:40:59.572213 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:41:00 crc kubenswrapper[4912]: I1203 01:41:00.673699 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"c4a43803629cefa948b3fe1050eb5e3b06fbca19c74ba5e14cc9709f2c863d56"} Dec 03 01:41:46 crc kubenswrapper[4912]: I1203 01:41:46.208839 4912 generic.go:334] "Generic (PLEG): container finished" podID="d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" containerID="0b9b1ed67406363c87d79c71afe55092a01400ad840bddbda9a7d26d18cfd111" exitCode=0 Dec 03 01:41:46 crc kubenswrapper[4912]: I1203 01:41:46.208892 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" event={"ID":"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e","Type":"ContainerDied","Data":"0b9b1ed67406363c87d79c71afe55092a01400ad840bddbda9a7d26d18cfd111"} Dec 03 01:41:47 crc kubenswrapper[4912]: I1203 01:41:47.931755 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028240 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph-nova-0\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028314 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-extra-config-0\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028370 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-inventory\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028452 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-0\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028512 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-1\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028538 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-0\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028574 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-custom-ceph-combined-ca-bundle\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028631 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-1\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028679 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028737 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ssh-key\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.028782 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj8pb\" (UniqueName: \"kubernetes.io/projected/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-kube-api-access-jj8pb\") pod \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\" (UID: \"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e\") " Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.087342 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.089517 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.089558 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph" (OuterVolumeSpecName: "ceph") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.090574 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-kube-api-access-jj8pb" (OuterVolumeSpecName: "kube-api-access-jj8pb") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "kube-api-access-jj8pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.099582 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-inventory" (OuterVolumeSpecName: "inventory") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.110136 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.130900 4912 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.130928 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.130939 4912 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.130953 4912 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.130966 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.130977 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj8pb\" (UniqueName: \"kubernetes.io/projected/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-kube-api-access-jj8pb\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.134580 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.136489 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.143690 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.143736 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.154218 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" (UID: "d65dbc95-9df7-49ec-9d0a-b883b70b4e3e"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.230525 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" event={"ID":"d65dbc95-9df7-49ec-9d0a-b883b70b4e3e","Type":"ContainerDied","Data":"71717653c7c0332bc90e2251e4596de2151d28db95e30c9f332306afb898ca3c"} Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.230566 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71717653c7c0332bc90e2251e4596de2151d28db95e30c9f332306afb898ca3c" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.230688 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.235786 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.235825 4912 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.235841 4912 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.235855 4912 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.235867 4912 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d65dbc95-9df7-49ec-9d0a-b883b70b4e3e-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.360145 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng"] Dec 03 01:41:48 crc kubenswrapper[4912]: E1203 01:41:48.360651 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.360677 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 03 01:41:48 crc kubenswrapper[4912]: E1203 01:41:48.360696 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerName="extract-content" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.360707 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerName="extract-content" Dec 03 01:41:48 crc kubenswrapper[4912]: E1203 01:41:48.360750 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerName="extract-utilities" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.360758 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerName="extract-utilities" Dec 03 01:41:48 crc kubenswrapper[4912]: E1203 01:41:48.360778 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerName="registry-server" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.360786 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerName="registry-server" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.361040 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d65dbc95-9df7-49ec-9d0a-b883b70b4e3e" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.361084 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9fb0b98-6f6f-432d-829e-80c2b0cf24fb" containerName="registry-server" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.362200 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.364295 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.364346 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.364376 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.365598 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.366069 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.367387 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.377012 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng"] Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.441254 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.441300 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.441332 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.441355 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.441384 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnz4t\" (UniqueName: \"kubernetes.io/projected/d6d6c301-1060-4d03-8e8e-6d776c912cc2-kube-api-access-jnz4t\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.441626 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.441755 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.441840 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.543492 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.543553 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.543800 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.543874 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.543924 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnz4t\" (UniqueName: \"kubernetes.io/projected/d6d6c301-1060-4d03-8e8e-6d776c912cc2-kube-api-access-jnz4t\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.543990 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.544042 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.544101 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.550473 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.550538 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.551208 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.551877 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.552360 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.552969 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.553060 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.564123 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnz4t\" (UniqueName: \"kubernetes.io/projected/d6d6c301-1060-4d03-8e8e-6d776c912cc2-kube-api-access-jnz4t\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:48 crc kubenswrapper[4912]: I1203 01:41:48.679556 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:41:49 crc kubenswrapper[4912]: I1203 01:41:49.273816 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng"] Dec 03 01:41:49 crc kubenswrapper[4912]: W1203 01:41:49.278585 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6d6c301_1060_4d03_8e8e_6d776c912cc2.slice/crio-13c9ac09f3e330cf073c34f17197b00568cabb0cd3add8bcd23ad755666cda97 WatchSource:0}: Error finding container 13c9ac09f3e330cf073c34f17197b00568cabb0cd3add8bcd23ad755666cda97: Status 404 returned error can't find the container with id 13c9ac09f3e330cf073c34f17197b00568cabb0cd3add8bcd23ad755666cda97 Dec 03 01:41:50 crc kubenswrapper[4912]: I1203 01:41:50.255155 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" event={"ID":"d6d6c301-1060-4d03-8e8e-6d776c912cc2","Type":"ContainerStarted","Data":"13c9ac09f3e330cf073c34f17197b00568cabb0cd3add8bcd23ad755666cda97"} Dec 03 01:41:51 crc kubenswrapper[4912]: I1203 01:41:51.277844 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" event={"ID":"d6d6c301-1060-4d03-8e8e-6d776c912cc2","Type":"ContainerStarted","Data":"775ce1e531b0441c98f9c130753144dba66e71cea4310393fd4858bd637d99e2"} Dec 03 01:41:51 crc kubenswrapper[4912]: I1203 01:41:51.316264 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" podStartSLOduration=2.7901192139999997 podStartE2EDuration="3.316232049s" podCreationTimestamp="2025-12-03 01:41:48 +0000 UTC" firstStartedPulling="2025-12-03 01:41:49.284112933 +0000 UTC m=+4694.926133513" lastFinishedPulling="2025-12-03 01:41:49.810225778 +0000 UTC m=+4695.452246348" observedRunningTime="2025-12-03 01:41:51.308978348 +0000 UTC m=+4696.950998978" watchObservedRunningTime="2025-12-03 01:41:51.316232049 +0000 UTC m=+4696.958252649" Dec 03 01:43:18 crc kubenswrapper[4912]: I1203 01:43:18.077843 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:43:18 crc kubenswrapper[4912]: I1203 01:43:18.078693 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:43:48 crc kubenswrapper[4912]: I1203 01:43:48.078418 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:43:48 crc kubenswrapper[4912]: I1203 01:43:48.079284 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.634834 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jkwfg"] Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.638963 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.679858 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkwfg"] Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.780454 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-utilities\") pod \"redhat-marketplace-jkwfg\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.780723 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vcmp\" (UniqueName: \"kubernetes.io/projected/f8ac981e-1f57-42ec-84c2-832ed4d602f7-kube-api-access-2vcmp\") pod \"redhat-marketplace-jkwfg\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.780794 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-catalog-content\") pod \"redhat-marketplace-jkwfg\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.882593 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-utilities\") pod \"redhat-marketplace-jkwfg\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.882687 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vcmp\" (UniqueName: \"kubernetes.io/projected/f8ac981e-1f57-42ec-84c2-832ed4d602f7-kube-api-access-2vcmp\") pod \"redhat-marketplace-jkwfg\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.882718 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-catalog-content\") pod \"redhat-marketplace-jkwfg\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.883164 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-catalog-content\") pod \"redhat-marketplace-jkwfg\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.883389 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-utilities\") pod \"redhat-marketplace-jkwfg\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.904754 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vcmp\" (UniqueName: \"kubernetes.io/projected/f8ac981e-1f57-42ec-84c2-832ed4d602f7-kube-api-access-2vcmp\") pod \"redhat-marketplace-jkwfg\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:08 crc kubenswrapper[4912]: I1203 01:44:08.965657 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:09 crc kubenswrapper[4912]: I1203 01:44:09.476921 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkwfg"] Dec 03 01:44:10 crc kubenswrapper[4912]: I1203 01:44:10.155270 4912 generic.go:334] "Generic (PLEG): container finished" podID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerID="418900c24fa3a0ab4d9e64996d3c5440ece50dcfefa0959f68c0e67d30d67e29" exitCode=0 Dec 03 01:44:10 crc kubenswrapper[4912]: I1203 01:44:10.155511 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkwfg" event={"ID":"f8ac981e-1f57-42ec-84c2-832ed4d602f7","Type":"ContainerDied","Data":"418900c24fa3a0ab4d9e64996d3c5440ece50dcfefa0959f68c0e67d30d67e29"} Dec 03 01:44:10 crc kubenswrapper[4912]: I1203 01:44:10.155577 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkwfg" event={"ID":"f8ac981e-1f57-42ec-84c2-832ed4d602f7","Type":"ContainerStarted","Data":"0437bd95f942542cfda83523e1e16179c9e59c9577ff46fcfdfb13a85458f79c"} Dec 03 01:44:10 crc kubenswrapper[4912]: I1203 01:44:10.157771 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:44:11 crc kubenswrapper[4912]: I1203 01:44:11.169654 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkwfg" event={"ID":"f8ac981e-1f57-42ec-84c2-832ed4d602f7","Type":"ContainerStarted","Data":"f0acc46d4aff74d0923e4c6ae5b08f397887d490ada9ef199f4750dbb19d5131"} Dec 03 01:44:12 crc kubenswrapper[4912]: I1203 01:44:12.183013 4912 generic.go:334] "Generic (PLEG): container finished" podID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerID="f0acc46d4aff74d0923e4c6ae5b08f397887d490ada9ef199f4750dbb19d5131" exitCode=0 Dec 03 01:44:12 crc kubenswrapper[4912]: I1203 01:44:12.183126 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkwfg" event={"ID":"f8ac981e-1f57-42ec-84c2-832ed4d602f7","Type":"ContainerDied","Data":"f0acc46d4aff74d0923e4c6ae5b08f397887d490ada9ef199f4750dbb19d5131"} Dec 03 01:44:13 crc kubenswrapper[4912]: I1203 01:44:13.202449 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkwfg" event={"ID":"f8ac981e-1f57-42ec-84c2-832ed4d602f7","Type":"ContainerStarted","Data":"43e745672b599bd726ae9ea29b2d12205a7a4feb89054c336a17c6dd275442c0"} Dec 03 01:44:13 crc kubenswrapper[4912]: I1203 01:44:13.222746 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jkwfg" podStartSLOduration=2.7384123689999997 podStartE2EDuration="5.222726611s" podCreationTimestamp="2025-12-03 01:44:08 +0000 UTC" firstStartedPulling="2025-12-03 01:44:10.157351703 +0000 UTC m=+4835.799372263" lastFinishedPulling="2025-12-03 01:44:12.641665945 +0000 UTC m=+4838.283686505" observedRunningTime="2025-12-03 01:44:13.220892543 +0000 UTC m=+4838.862913113" watchObservedRunningTime="2025-12-03 01:44:13.222726611 +0000 UTC m=+4838.864747181" Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.077854 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.078486 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.078547 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.079818 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c4a43803629cefa948b3fe1050eb5e3b06fbca19c74ba5e14cc9709f2c863d56"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.079957 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://c4a43803629cefa948b3fe1050eb5e3b06fbca19c74ba5e14cc9709f2c863d56" gracePeriod=600 Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.278639 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="c4a43803629cefa948b3fe1050eb5e3b06fbca19c74ba5e14cc9709f2c863d56" exitCode=0 Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.278724 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"c4a43803629cefa948b3fe1050eb5e3b06fbca19c74ba5e14cc9709f2c863d56"} Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.278957 4912 scope.go:117] "RemoveContainer" containerID="839c84e4314c7b18278969316c7fe81a37aba7bf778e9fa229ad6c802b1bbf00" Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.965886 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:18 crc kubenswrapper[4912]: I1203 01:44:18.966298 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.050583 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.297742 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268"} Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.379555 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vj962"] Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.383012 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.389786 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vj962"] Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.402498 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.433770 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-catalog-content\") pod \"redhat-operators-vj962\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.433823 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhd2n\" (UniqueName: \"kubernetes.io/projected/410c1ea5-8979-4769-bba7-6b5510ee59c8-kube-api-access-qhd2n\") pod \"redhat-operators-vj962\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.433956 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-utilities\") pod \"redhat-operators-vj962\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.536213 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-utilities\") pod \"redhat-operators-vj962\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.536401 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-catalog-content\") pod \"redhat-operators-vj962\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.536447 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhd2n\" (UniqueName: \"kubernetes.io/projected/410c1ea5-8979-4769-bba7-6b5510ee59c8-kube-api-access-qhd2n\") pod \"redhat-operators-vj962\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.537279 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-utilities\") pod \"redhat-operators-vj962\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.537517 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-catalog-content\") pod \"redhat-operators-vj962\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.559710 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhd2n\" (UniqueName: \"kubernetes.io/projected/410c1ea5-8979-4769-bba7-6b5510ee59c8-kube-api-access-qhd2n\") pod \"redhat-operators-vj962\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:19 crc kubenswrapper[4912]: I1203 01:44:19.705936 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:20 crc kubenswrapper[4912]: I1203 01:44:20.016664 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vj962"] Dec 03 01:44:20 crc kubenswrapper[4912]: W1203 01:44:20.028741 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod410c1ea5_8979_4769_bba7_6b5510ee59c8.slice/crio-d5516f6912e6f1745adcda6abb65e51d7793f19b5a6c7626d0739a1d75d802a6 WatchSource:0}: Error finding container d5516f6912e6f1745adcda6abb65e51d7793f19b5a6c7626d0739a1d75d802a6: Status 404 returned error can't find the container with id d5516f6912e6f1745adcda6abb65e51d7793f19b5a6c7626d0739a1d75d802a6 Dec 03 01:44:20 crc kubenswrapper[4912]: I1203 01:44:20.313238 4912 generic.go:334] "Generic (PLEG): container finished" podID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerID="62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648" exitCode=0 Dec 03 01:44:20 crc kubenswrapper[4912]: I1203 01:44:20.313298 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj962" event={"ID":"410c1ea5-8979-4769-bba7-6b5510ee59c8","Type":"ContainerDied","Data":"62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648"} Dec 03 01:44:20 crc kubenswrapper[4912]: I1203 01:44:20.313335 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj962" event={"ID":"410c1ea5-8979-4769-bba7-6b5510ee59c8","Type":"ContainerStarted","Data":"d5516f6912e6f1745adcda6abb65e51d7793f19b5a6c7626d0739a1d75d802a6"} Dec 03 01:44:21 crc kubenswrapper[4912]: I1203 01:44:21.328225 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj962" event={"ID":"410c1ea5-8979-4769-bba7-6b5510ee59c8","Type":"ContainerStarted","Data":"752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684"} Dec 03 01:44:21 crc kubenswrapper[4912]: I1203 01:44:21.700958 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkwfg"] Dec 03 01:44:21 crc kubenswrapper[4912]: I1203 01:44:21.701254 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jkwfg" podUID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerName="registry-server" containerID="cri-o://43e745672b599bd726ae9ea29b2d12205a7a4feb89054c336a17c6dd275442c0" gracePeriod=2 Dec 03 01:44:23 crc kubenswrapper[4912]: I1203 01:44:23.355808 4912 generic.go:334] "Generic (PLEG): container finished" podID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerID="752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684" exitCode=0 Dec 03 01:44:23 crc kubenswrapper[4912]: I1203 01:44:23.355870 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj962" event={"ID":"410c1ea5-8979-4769-bba7-6b5510ee59c8","Type":"ContainerDied","Data":"752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684"} Dec 03 01:44:23 crc kubenswrapper[4912]: I1203 01:44:23.361866 4912 generic.go:334] "Generic (PLEG): container finished" podID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerID="43e745672b599bd726ae9ea29b2d12205a7a4feb89054c336a17c6dd275442c0" exitCode=0 Dec 03 01:44:23 crc kubenswrapper[4912]: I1203 01:44:23.361906 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkwfg" event={"ID":"f8ac981e-1f57-42ec-84c2-832ed4d602f7","Type":"ContainerDied","Data":"43e745672b599bd726ae9ea29b2d12205a7a4feb89054c336a17c6dd275442c0"} Dec 03 01:44:23 crc kubenswrapper[4912]: E1203 01:44:23.423872 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod410c1ea5_8979_4769_bba7_6b5510ee59c8.slice/crio-conmon-752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684.scope\": RecentStats: unable to find data in memory cache]" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.146769 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.243414 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-utilities\") pod \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.243510 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vcmp\" (UniqueName: \"kubernetes.io/projected/f8ac981e-1f57-42ec-84c2-832ed4d602f7-kube-api-access-2vcmp\") pod \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.243687 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-catalog-content\") pod \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\" (UID: \"f8ac981e-1f57-42ec-84c2-832ed4d602f7\") " Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.249461 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-utilities" (OuterVolumeSpecName: "utilities") pod "f8ac981e-1f57-42ec-84c2-832ed4d602f7" (UID: "f8ac981e-1f57-42ec-84c2-832ed4d602f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.249615 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8ac981e-1f57-42ec-84c2-832ed4d602f7-kube-api-access-2vcmp" (OuterVolumeSpecName: "kube-api-access-2vcmp") pod "f8ac981e-1f57-42ec-84c2-832ed4d602f7" (UID: "f8ac981e-1f57-42ec-84c2-832ed4d602f7"). InnerVolumeSpecName "kube-api-access-2vcmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.276876 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8ac981e-1f57-42ec-84c2-832ed4d602f7" (UID: "f8ac981e-1f57-42ec-84c2-832ed4d602f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.346301 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.346346 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8ac981e-1f57-42ec-84c2-832ed4d602f7-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.346364 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vcmp\" (UniqueName: \"kubernetes.io/projected/f8ac981e-1f57-42ec-84c2-832ed4d602f7-kube-api-access-2vcmp\") on node \"crc\" DevicePath \"\"" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.377004 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj962" event={"ID":"410c1ea5-8979-4769-bba7-6b5510ee59c8","Type":"ContainerStarted","Data":"e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23"} Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.382175 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jkwfg" event={"ID":"f8ac981e-1f57-42ec-84c2-832ed4d602f7","Type":"ContainerDied","Data":"0437bd95f942542cfda83523e1e16179c9e59c9577ff46fcfdfb13a85458f79c"} Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.382235 4912 scope.go:117] "RemoveContainer" containerID="43e745672b599bd726ae9ea29b2d12205a7a4feb89054c336a17c6dd275442c0" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.382281 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jkwfg" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.406711 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vj962" podStartSLOduration=1.928316854 podStartE2EDuration="5.406684044s" podCreationTimestamp="2025-12-03 01:44:19 +0000 UTC" firstStartedPulling="2025-12-03 01:44:20.315827589 +0000 UTC m=+4845.957848149" lastFinishedPulling="2025-12-03 01:44:23.794194769 +0000 UTC m=+4849.436215339" observedRunningTime="2025-12-03 01:44:24.395784777 +0000 UTC m=+4850.037805347" watchObservedRunningTime="2025-12-03 01:44:24.406684044 +0000 UTC m=+4850.048704614" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.438208 4912 scope.go:117] "RemoveContainer" containerID="f0acc46d4aff74d0923e4c6ae5b08f397887d490ada9ef199f4750dbb19d5131" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.455694 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkwfg"] Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.471264 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jkwfg"] Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.474105 4912 scope.go:117] "RemoveContainer" containerID="418900c24fa3a0ab4d9e64996d3c5440ece50dcfefa0959f68c0e67d30d67e29" Dec 03 01:44:24 crc kubenswrapper[4912]: I1203 01:44:24.588804 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" path="/var/lib/kubelet/pods/f8ac981e-1f57-42ec-84c2-832ed4d602f7/volumes" Dec 03 01:44:29 crc kubenswrapper[4912]: I1203 01:44:29.706574 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:29 crc kubenswrapper[4912]: I1203 01:44:29.706976 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:30 crc kubenswrapper[4912]: I1203 01:44:30.765895 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vj962" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerName="registry-server" probeResult="failure" output=< Dec 03 01:44:30 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 01:44:30 crc kubenswrapper[4912]: > Dec 03 01:44:39 crc kubenswrapper[4912]: I1203 01:44:39.770699 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:39 crc kubenswrapper[4912]: I1203 01:44:39.846346 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:40 crc kubenswrapper[4912]: I1203 01:44:40.030884 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vj962"] Dec 03 01:44:41 crc kubenswrapper[4912]: I1203 01:44:41.604688 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vj962" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerName="registry-server" containerID="cri-o://e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23" gracePeriod=2 Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.238409 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.346301 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhd2n\" (UniqueName: \"kubernetes.io/projected/410c1ea5-8979-4769-bba7-6b5510ee59c8-kube-api-access-qhd2n\") pod \"410c1ea5-8979-4769-bba7-6b5510ee59c8\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.346415 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-catalog-content\") pod \"410c1ea5-8979-4769-bba7-6b5510ee59c8\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.346618 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-utilities\") pod \"410c1ea5-8979-4769-bba7-6b5510ee59c8\" (UID: \"410c1ea5-8979-4769-bba7-6b5510ee59c8\") " Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.347648 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-utilities" (OuterVolumeSpecName: "utilities") pod "410c1ea5-8979-4769-bba7-6b5510ee59c8" (UID: "410c1ea5-8979-4769-bba7-6b5510ee59c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.352097 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/410c1ea5-8979-4769-bba7-6b5510ee59c8-kube-api-access-qhd2n" (OuterVolumeSpecName: "kube-api-access-qhd2n") pod "410c1ea5-8979-4769-bba7-6b5510ee59c8" (UID: "410c1ea5-8979-4769-bba7-6b5510ee59c8"). InnerVolumeSpecName "kube-api-access-qhd2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.449407 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhd2n\" (UniqueName: \"kubernetes.io/projected/410c1ea5-8979-4769-bba7-6b5510ee59c8-kube-api-access-qhd2n\") on node \"crc\" DevicePath \"\"" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.449469 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.495789 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "410c1ea5-8979-4769-bba7-6b5510ee59c8" (UID: "410c1ea5-8979-4769-bba7-6b5510ee59c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.551771 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/410c1ea5-8979-4769-bba7-6b5510ee59c8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.630455 4912 generic.go:334] "Generic (PLEG): container finished" podID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerID="e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23" exitCode=0 Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.630510 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj962" event={"ID":"410c1ea5-8979-4769-bba7-6b5510ee59c8","Type":"ContainerDied","Data":"e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23"} Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.630542 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj962" event={"ID":"410c1ea5-8979-4769-bba7-6b5510ee59c8","Type":"ContainerDied","Data":"d5516f6912e6f1745adcda6abb65e51d7793f19b5a6c7626d0739a1d75d802a6"} Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.630561 4912 scope.go:117] "RemoveContainer" containerID="e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.630822 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj962" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.659501 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vj962"] Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.660941 4912 scope.go:117] "RemoveContainer" containerID="752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.669421 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vj962"] Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.691711 4912 scope.go:117] "RemoveContainer" containerID="62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.739506 4912 scope.go:117] "RemoveContainer" containerID="e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23" Dec 03 01:44:42 crc kubenswrapper[4912]: E1203 01:44:42.739978 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23\": container with ID starting with e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23 not found: ID does not exist" containerID="e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.740093 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23"} err="failed to get container status \"e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23\": rpc error: code = NotFound desc = could not find container \"e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23\": container with ID starting with e855abd15c60d10e0a6f864207b63a5bea3ebe586b88dd9835c7c832d3515c23 not found: ID does not exist" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.740178 4912 scope.go:117] "RemoveContainer" containerID="752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684" Dec 03 01:44:42 crc kubenswrapper[4912]: E1203 01:44:42.740580 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684\": container with ID starting with 752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684 not found: ID does not exist" containerID="752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.740624 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684"} err="failed to get container status \"752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684\": rpc error: code = NotFound desc = could not find container \"752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684\": container with ID starting with 752084c47546a5d9e7aceae18b4e17fb89b732128977ff4cb0d1fe3b01912684 not found: ID does not exist" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.740654 4912 scope.go:117] "RemoveContainer" containerID="62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648" Dec 03 01:44:42 crc kubenswrapper[4912]: E1203 01:44:42.740950 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648\": container with ID starting with 62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648 not found: ID does not exist" containerID="62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648" Dec 03 01:44:42 crc kubenswrapper[4912]: I1203 01:44:42.740981 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648"} err="failed to get container status \"62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648\": rpc error: code = NotFound desc = could not find container \"62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648\": container with ID starting with 62bed278ad44780c491ea42c56da81f29af7950574d797808166c2a0537ff648 not found: ID does not exist" Dec 03 01:44:44 crc kubenswrapper[4912]: I1203 01:44:44.591915 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" path="/var/lib/kubelet/pods/410c1ea5-8979-4769-bba7-6b5510ee59c8/volumes" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.159648 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf"] Dec 03 01:45:00 crc kubenswrapper[4912]: E1203 01:45:00.160843 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerName="registry-server" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.160865 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerName="registry-server" Dec 03 01:45:00 crc kubenswrapper[4912]: E1203 01:45:00.160898 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerName="extract-content" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.160909 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerName="extract-content" Dec 03 01:45:00 crc kubenswrapper[4912]: E1203 01:45:00.160935 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerName="extract-utilities" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.160945 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerName="extract-utilities" Dec 03 01:45:00 crc kubenswrapper[4912]: E1203 01:45:00.160967 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerName="extract-content" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.160974 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerName="extract-content" Dec 03 01:45:00 crc kubenswrapper[4912]: E1203 01:45:00.160998 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerName="registry-server" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.161006 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerName="registry-server" Dec 03 01:45:00 crc kubenswrapper[4912]: E1203 01:45:00.161033 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerName="extract-utilities" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.161042 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerName="extract-utilities" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.161347 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8ac981e-1f57-42ec-84c2-832ed4d602f7" containerName="registry-server" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.161363 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="410c1ea5-8979-4769-bba7-6b5510ee59c8" containerName="registry-server" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.162419 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.164883 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.164957 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.172884 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf"] Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.291910 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-config-volume\") pod \"collect-profiles-29412105-jbkkf\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.292028 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-secret-volume\") pod \"collect-profiles-29412105-jbkkf\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.292127 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fh96\" (UniqueName: \"kubernetes.io/projected/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-kube-api-access-9fh96\") pod \"collect-profiles-29412105-jbkkf\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.394146 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-secret-volume\") pod \"collect-profiles-29412105-jbkkf\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.394527 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fh96\" (UniqueName: \"kubernetes.io/projected/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-kube-api-access-9fh96\") pod \"collect-profiles-29412105-jbkkf\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.394627 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-config-volume\") pod \"collect-profiles-29412105-jbkkf\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.395486 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-config-volume\") pod \"collect-profiles-29412105-jbkkf\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.402225 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-secret-volume\") pod \"collect-profiles-29412105-jbkkf\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.410571 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fh96\" (UniqueName: \"kubernetes.io/projected/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-kube-api-access-9fh96\") pod \"collect-profiles-29412105-jbkkf\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.486062 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:00 crc kubenswrapper[4912]: I1203 01:45:00.980859 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf"] Dec 03 01:45:01 crc kubenswrapper[4912]: I1203 01:45:01.862555 4912 generic.go:334] "Generic (PLEG): container finished" podID="9dff648e-50e3-433c-b9e6-d3a83c0ef3aa" containerID="c114602a364bed6f635a573ee34f2bb3490154855fa8b6090cd3adf54bf8e76a" exitCode=0 Dec 03 01:45:01 crc kubenswrapper[4912]: I1203 01:45:01.862712 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" event={"ID":"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa","Type":"ContainerDied","Data":"c114602a364bed6f635a573ee34f2bb3490154855fa8b6090cd3adf54bf8e76a"} Dec 03 01:45:01 crc kubenswrapper[4912]: I1203 01:45:01.862936 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" event={"ID":"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa","Type":"ContainerStarted","Data":"c789b8465a468ad436f284802ae54c8ae5ca94b0f10f5773f7ca7482a531e33c"} Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.318851 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.458099 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fh96\" (UniqueName: \"kubernetes.io/projected/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-kube-api-access-9fh96\") pod \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.458304 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-secret-volume\") pod \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.458348 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-config-volume\") pod \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\" (UID: \"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa\") " Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.459349 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-config-volume" (OuterVolumeSpecName: "config-volume") pod "9dff648e-50e3-433c-b9e6-d3a83c0ef3aa" (UID: "9dff648e-50e3-433c-b9e6-d3a83c0ef3aa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.463577 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9dff648e-50e3-433c-b9e6-d3a83c0ef3aa" (UID: "9dff648e-50e3-433c-b9e6-d3a83c0ef3aa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.465949 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-kube-api-access-9fh96" (OuterVolumeSpecName: "kube-api-access-9fh96") pod "9dff648e-50e3-433c-b9e6-d3a83c0ef3aa" (UID: "9dff648e-50e3-433c-b9e6-d3a83c0ef3aa"). InnerVolumeSpecName "kube-api-access-9fh96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.560801 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fh96\" (UniqueName: \"kubernetes.io/projected/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-kube-api-access-9fh96\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.560838 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.560849 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.880469 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" event={"ID":"9dff648e-50e3-433c-b9e6-d3a83c0ef3aa","Type":"ContainerDied","Data":"c789b8465a468ad436f284802ae54c8ae5ca94b0f10f5773f7ca7482a531e33c"} Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.880527 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c789b8465a468ad436f284802ae54c8ae5ca94b0f10f5773f7ca7482a531e33c" Dec 03 01:45:03 crc kubenswrapper[4912]: I1203 01:45:03.880531 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf" Dec 03 01:45:04 crc kubenswrapper[4912]: I1203 01:45:04.399737 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6"] Dec 03 01:45:04 crc kubenswrapper[4912]: I1203 01:45:04.408861 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412060-ktkz6"] Dec 03 01:45:04 crc kubenswrapper[4912]: I1203 01:45:04.587222 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f4130dc-bed8-4576-a343-2a41db43d477" path="/var/lib/kubelet/pods/4f4130dc-bed8-4576-a343-2a41db43d477/volumes" Dec 03 01:45:26 crc kubenswrapper[4912]: I1203 01:45:26.114365 4912 scope.go:117] "RemoveContainer" containerID="96a493340f606df04130503787a40dc53f344b6ebff9ef51e2c604722d3e532a" Dec 03 01:45:26 crc kubenswrapper[4912]: I1203 01:45:26.157383 4912 generic.go:334] "Generic (PLEG): container finished" podID="d6d6c301-1060-4d03-8e8e-6d776c912cc2" containerID="775ce1e531b0441c98f9c130753144dba66e71cea4310393fd4858bd637d99e2" exitCode=0 Dec 03 01:45:26 crc kubenswrapper[4912]: I1203 01:45:26.157465 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" event={"ID":"d6d6c301-1060-4d03-8e8e-6d776c912cc2","Type":"ContainerDied","Data":"775ce1e531b0441c98f9c130753144dba66e71cea4310393fd4858bd637d99e2"} Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.779575 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.858574 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-0\") pod \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.858681 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceph\") pod \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.858719 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnz4t\" (UniqueName: \"kubernetes.io/projected/d6d6c301-1060-4d03-8e8e-6d776c912cc2-kube-api-access-jnz4t\") pod \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.858751 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-inventory\") pod \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.858783 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-telemetry-combined-ca-bundle\") pod \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.858826 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-1\") pod \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.859608 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-2\") pod \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.859653 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ssh-key\") pod \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\" (UID: \"d6d6c301-1060-4d03-8e8e-6d776c912cc2\") " Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.865476 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceph" (OuterVolumeSpecName: "ceph") pod "d6d6c301-1060-4d03-8e8e-6d776c912cc2" (UID: "d6d6c301-1060-4d03-8e8e-6d776c912cc2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.878242 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6d6c301-1060-4d03-8e8e-6d776c912cc2-kube-api-access-jnz4t" (OuterVolumeSpecName: "kube-api-access-jnz4t") pod "d6d6c301-1060-4d03-8e8e-6d776c912cc2" (UID: "d6d6c301-1060-4d03-8e8e-6d776c912cc2"). InnerVolumeSpecName "kube-api-access-jnz4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.883422 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d6d6c301-1060-4d03-8e8e-6d776c912cc2" (UID: "d6d6c301-1060-4d03-8e8e-6d776c912cc2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.897118 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d6d6c301-1060-4d03-8e8e-6d776c912cc2" (UID: "d6d6c301-1060-4d03-8e8e-6d776c912cc2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.902626 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "d6d6c301-1060-4d03-8e8e-6d776c912cc2" (UID: "d6d6c301-1060-4d03-8e8e-6d776c912cc2"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.903513 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "d6d6c301-1060-4d03-8e8e-6d776c912cc2" (UID: "d6d6c301-1060-4d03-8e8e-6d776c912cc2"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.907297 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "d6d6c301-1060-4d03-8e8e-6d776c912cc2" (UID: "d6d6c301-1060-4d03-8e8e-6d776c912cc2"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.914051 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-inventory" (OuterVolumeSpecName: "inventory") pod "d6d6c301-1060-4d03-8e8e-6d776c912cc2" (UID: "d6d6c301-1060-4d03-8e8e-6d776c912cc2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.962483 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.962529 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.962544 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.962557 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.962575 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.962592 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnz4t\" (UniqueName: \"kubernetes.io/projected/d6d6c301-1060-4d03-8e8e-6d776c912cc2-kube-api-access-jnz4t\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.962608 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:27 crc kubenswrapper[4912]: I1203 01:45:27.962623 4912 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6d6c301-1060-4d03-8e8e-6d776c912cc2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.186890 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" event={"ID":"d6d6c301-1060-4d03-8e8e-6d776c912cc2","Type":"ContainerDied","Data":"13c9ac09f3e330cf073c34f17197b00568cabb0cd3add8bcd23ad755666cda97"} Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.187328 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13c9ac09f3e330cf073c34f17197b00568cabb0cd3add8bcd23ad755666cda97" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.187099 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.343406 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm"] Dec 03 01:45:28 crc kubenswrapper[4912]: E1203 01:45:28.344109 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dff648e-50e3-433c-b9e6-d3a83c0ef3aa" containerName="collect-profiles" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.344142 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dff648e-50e3-433c-b9e6-d3a83c0ef3aa" containerName="collect-profiles" Dec 03 01:45:28 crc kubenswrapper[4912]: E1203 01:45:28.344170 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6d6c301-1060-4d03-8e8e-6d776c912cc2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.344183 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6d6c301-1060-4d03-8e8e-6d776c912cc2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.344595 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6d6c301-1060-4d03-8e8e-6d776c912cc2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.344625 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dff648e-50e3-433c-b9e6-d3a83c0ef3aa" containerName="collect-profiles" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.345812 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.350735 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.350949 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.351179 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.351498 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.351669 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.352754 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.362927 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm"] Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.474376 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.474523 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rxnj\" (UniqueName: \"kubernetes.io/projected/e34cc78e-e946-4d45-a3da-962e308ad972-kube-api-access-4rxnj\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.474564 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.474608 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.474646 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.474842 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.474958 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.475000 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.577705 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.577868 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.577942 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.577997 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.578212 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.578316 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rxnj\" (UniqueName: \"kubernetes.io/projected/e34cc78e-e946-4d45-a3da-962e308ad972-kube-api-access-4rxnj\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.578363 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.578416 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.583289 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.583370 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.584043 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.584695 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.585386 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.585834 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.593988 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.598992 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rxnj\" (UniqueName: \"kubernetes.io/projected/e34cc78e-e946-4d45-a3da-962e308ad972-kube-api-access-4rxnj\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:28 crc kubenswrapper[4912]: I1203 01:45:28.665373 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:45:29 crc kubenswrapper[4912]: I1203 01:45:29.230892 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm"] Dec 03 01:45:29 crc kubenswrapper[4912]: W1203 01:45:29.233760 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode34cc78e_e946_4d45_a3da_962e308ad972.slice/crio-6193ce126c62ff3b76461baafcdc616bf4d6186a19c6bf34efdd9feb287ec8c9 WatchSource:0}: Error finding container 6193ce126c62ff3b76461baafcdc616bf4d6186a19c6bf34efdd9feb287ec8c9: Status 404 returned error can't find the container with id 6193ce126c62ff3b76461baafcdc616bf4d6186a19c6bf34efdd9feb287ec8c9 Dec 03 01:45:30 crc kubenswrapper[4912]: I1203 01:45:30.207027 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" event={"ID":"e34cc78e-e946-4d45-a3da-962e308ad972","Type":"ContainerStarted","Data":"783e45d02e45fb0e26984a31367a4fb43eb798bd7b52295c82eff5708cbb9d9c"} Dec 03 01:45:30 crc kubenswrapper[4912]: I1203 01:45:30.207288 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" event={"ID":"e34cc78e-e946-4d45-a3da-962e308ad972","Type":"ContainerStarted","Data":"6193ce126c62ff3b76461baafcdc616bf4d6186a19c6bf34efdd9feb287ec8c9"} Dec 03 01:45:30 crc kubenswrapper[4912]: I1203 01:45:30.236252 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" podStartSLOduration=1.56627692 podStartE2EDuration="2.236227602s" podCreationTimestamp="2025-12-03 01:45:28 +0000 UTC" firstStartedPulling="2025-12-03 01:45:29.236985129 +0000 UTC m=+4914.879005699" lastFinishedPulling="2025-12-03 01:45:29.906935811 +0000 UTC m=+4915.548956381" observedRunningTime="2025-12-03 01:45:30.229954237 +0000 UTC m=+4915.871974797" watchObservedRunningTime="2025-12-03 01:45:30.236227602 +0000 UTC m=+4915.878248172" Dec 03 01:46:18 crc kubenswrapper[4912]: I1203 01:46:18.078385 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:46:18 crc kubenswrapper[4912]: I1203 01:46:18.079074 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:46:48 crc kubenswrapper[4912]: I1203 01:46:48.077615 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:46:48 crc kubenswrapper[4912]: I1203 01:46:48.078304 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:47:18 crc kubenswrapper[4912]: I1203 01:47:18.077402 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:47:18 crc kubenswrapper[4912]: I1203 01:47:18.079513 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:47:18 crc kubenswrapper[4912]: I1203 01:47:18.079784 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:47:18 crc kubenswrapper[4912]: I1203 01:47:18.080928 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:47:18 crc kubenswrapper[4912]: I1203 01:47:18.081122 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" gracePeriod=600 Dec 03 01:47:18 crc kubenswrapper[4912]: E1203 01:47:18.204145 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:47:18 crc kubenswrapper[4912]: I1203 01:47:18.551198 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" exitCode=0 Dec 03 01:47:18 crc kubenswrapper[4912]: I1203 01:47:18.551286 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268"} Dec 03 01:47:18 crc kubenswrapper[4912]: I1203 01:47:18.551356 4912 scope.go:117] "RemoveContainer" containerID="c4a43803629cefa948b3fe1050eb5e3b06fbca19c74ba5e14cc9709f2c863d56" Dec 03 01:47:18 crc kubenswrapper[4912]: I1203 01:47:18.552662 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:47:18 crc kubenswrapper[4912]: E1203 01:47:18.553280 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:47:30 crc kubenswrapper[4912]: I1203 01:47:30.572184 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:47:30 crc kubenswrapper[4912]: E1203 01:47:30.573198 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:47:45 crc kubenswrapper[4912]: I1203 01:47:45.572598 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:47:45 crc kubenswrapper[4912]: E1203 01:47:45.576086 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:47:57 crc kubenswrapper[4912]: I1203 01:47:57.572200 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:47:57 crc kubenswrapper[4912]: E1203 01:47:57.573391 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:48:09 crc kubenswrapper[4912]: I1203 01:48:09.774335 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cqmld"] Dec 03 01:48:09 crc kubenswrapper[4912]: I1203 01:48:09.781054 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:09 crc kubenswrapper[4912]: I1203 01:48:09.794681 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cqmld"] Dec 03 01:48:09 crc kubenswrapper[4912]: I1203 01:48:09.895400 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpcct\" (UniqueName: \"kubernetes.io/projected/866c1e6f-1d40-4706-8a0b-e95259e192ba-kube-api-access-dpcct\") pod \"certified-operators-cqmld\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:09 crc kubenswrapper[4912]: I1203 01:48:09.895476 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-catalog-content\") pod \"certified-operators-cqmld\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:09 crc kubenswrapper[4912]: I1203 01:48:09.895531 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-utilities\") pod \"certified-operators-cqmld\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:10 crc kubenswrapper[4912]: I1203 01:48:10.001720 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpcct\" (UniqueName: \"kubernetes.io/projected/866c1e6f-1d40-4706-8a0b-e95259e192ba-kube-api-access-dpcct\") pod \"certified-operators-cqmld\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:10 crc kubenswrapper[4912]: I1203 01:48:10.001999 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-catalog-content\") pod \"certified-operators-cqmld\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:10 crc kubenswrapper[4912]: I1203 01:48:10.002075 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-utilities\") pod \"certified-operators-cqmld\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:10 crc kubenswrapper[4912]: I1203 01:48:10.003123 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-utilities\") pod \"certified-operators-cqmld\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:10 crc kubenswrapper[4912]: I1203 01:48:10.003376 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-catalog-content\") pod \"certified-operators-cqmld\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:10 crc kubenswrapper[4912]: I1203 01:48:10.025310 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpcct\" (UniqueName: \"kubernetes.io/projected/866c1e6f-1d40-4706-8a0b-e95259e192ba-kube-api-access-dpcct\") pod \"certified-operators-cqmld\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:10 crc kubenswrapper[4912]: I1203 01:48:10.156467 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:10 crc kubenswrapper[4912]: I1203 01:48:10.571665 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:48:10 crc kubenswrapper[4912]: E1203 01:48:10.572170 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:48:10 crc kubenswrapper[4912]: I1203 01:48:10.716634 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cqmld"] Dec 03 01:48:11 crc kubenswrapper[4912]: I1203 01:48:11.309443 4912 generic.go:334] "Generic (PLEG): container finished" podID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerID="046a5a967b2d34f4bdd2f85f053a072a2ef0de3c04c6cdcaf59a6e5f34610bc9" exitCode=0 Dec 03 01:48:11 crc kubenswrapper[4912]: I1203 01:48:11.309550 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqmld" event={"ID":"866c1e6f-1d40-4706-8a0b-e95259e192ba","Type":"ContainerDied","Data":"046a5a967b2d34f4bdd2f85f053a072a2ef0de3c04c6cdcaf59a6e5f34610bc9"} Dec 03 01:48:11 crc kubenswrapper[4912]: I1203 01:48:11.309993 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqmld" event={"ID":"866c1e6f-1d40-4706-8a0b-e95259e192ba","Type":"ContainerStarted","Data":"6148334d61c5ab8e29dda7d64f96e77628234ec8c123304c33e4d4513ac9dfbc"} Dec 03 01:48:12 crc kubenswrapper[4912]: I1203 01:48:12.326131 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqmld" event={"ID":"866c1e6f-1d40-4706-8a0b-e95259e192ba","Type":"ContainerStarted","Data":"cd8e0347ad84a6fc18c9b73f8819b9dc6739c8f29be8cdc0d3f724fb37e6143f"} Dec 03 01:48:13 crc kubenswrapper[4912]: I1203 01:48:13.347030 4912 generic.go:334] "Generic (PLEG): container finished" podID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerID="cd8e0347ad84a6fc18c9b73f8819b9dc6739c8f29be8cdc0d3f724fb37e6143f" exitCode=0 Dec 03 01:48:13 crc kubenswrapper[4912]: I1203 01:48:13.347109 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqmld" event={"ID":"866c1e6f-1d40-4706-8a0b-e95259e192ba","Type":"ContainerDied","Data":"cd8e0347ad84a6fc18c9b73f8819b9dc6739c8f29be8cdc0d3f724fb37e6143f"} Dec 03 01:48:14 crc kubenswrapper[4912]: I1203 01:48:14.362492 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqmld" event={"ID":"866c1e6f-1d40-4706-8a0b-e95259e192ba","Type":"ContainerStarted","Data":"7afd0a9e398bb78fed3b7bdda38f9f27420d7c65264353e7634bedd6e408ee14"} Dec 03 01:48:14 crc kubenswrapper[4912]: I1203 01:48:14.392202 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cqmld" podStartSLOduration=2.914559958 podStartE2EDuration="5.392179289s" podCreationTimestamp="2025-12-03 01:48:09 +0000 UTC" firstStartedPulling="2025-12-03 01:48:11.313096491 +0000 UTC m=+5076.955117051" lastFinishedPulling="2025-12-03 01:48:13.790715812 +0000 UTC m=+5079.432736382" observedRunningTime="2025-12-03 01:48:14.386863866 +0000 UTC m=+5080.028884436" watchObservedRunningTime="2025-12-03 01:48:14.392179289 +0000 UTC m=+5080.034199859" Dec 03 01:48:20 crc kubenswrapper[4912]: I1203 01:48:20.156614 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:20 crc kubenswrapper[4912]: I1203 01:48:20.157157 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:20 crc kubenswrapper[4912]: I1203 01:48:20.204722 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:20 crc kubenswrapper[4912]: I1203 01:48:20.507177 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:20 crc kubenswrapper[4912]: I1203 01:48:20.566128 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cqmld"] Dec 03 01:48:22 crc kubenswrapper[4912]: I1203 01:48:22.467814 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cqmld" podUID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerName="registry-server" containerID="cri-o://7afd0a9e398bb78fed3b7bdda38f9f27420d7c65264353e7634bedd6e408ee14" gracePeriod=2 Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.486620 4912 generic.go:334] "Generic (PLEG): container finished" podID="e34cc78e-e946-4d45-a3da-962e308ad972" containerID="783e45d02e45fb0e26984a31367a4fb43eb798bd7b52295c82eff5708cbb9d9c" exitCode=0 Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.486685 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" event={"ID":"e34cc78e-e946-4d45-a3da-962e308ad972","Type":"ContainerDied","Data":"783e45d02e45fb0e26984a31367a4fb43eb798bd7b52295c82eff5708cbb9d9c"} Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.491174 4912 generic.go:334] "Generic (PLEG): container finished" podID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerID="7afd0a9e398bb78fed3b7bdda38f9f27420d7c65264353e7634bedd6e408ee14" exitCode=0 Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.491255 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqmld" event={"ID":"866c1e6f-1d40-4706-8a0b-e95259e192ba","Type":"ContainerDied","Data":"7afd0a9e398bb78fed3b7bdda38f9f27420d7c65264353e7634bedd6e408ee14"} Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.491678 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cqmld" event={"ID":"866c1e6f-1d40-4706-8a0b-e95259e192ba","Type":"ContainerDied","Data":"6148334d61c5ab8e29dda7d64f96e77628234ec8c123304c33e4d4513ac9dfbc"} Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.491863 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6148334d61c5ab8e29dda7d64f96e77628234ec8c123304c33e4d4513ac9dfbc" Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.571626 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:48:23 crc kubenswrapper[4912]: E1203 01:48:23.572515 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.582001 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.763162 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-utilities\") pod \"866c1e6f-1d40-4706-8a0b-e95259e192ba\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.763308 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpcct\" (UniqueName: \"kubernetes.io/projected/866c1e6f-1d40-4706-8a0b-e95259e192ba-kube-api-access-dpcct\") pod \"866c1e6f-1d40-4706-8a0b-e95259e192ba\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.763390 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-catalog-content\") pod \"866c1e6f-1d40-4706-8a0b-e95259e192ba\" (UID: \"866c1e6f-1d40-4706-8a0b-e95259e192ba\") " Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.764329 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-utilities" (OuterVolumeSpecName: "utilities") pod "866c1e6f-1d40-4706-8a0b-e95259e192ba" (UID: "866c1e6f-1d40-4706-8a0b-e95259e192ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.771763 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/866c1e6f-1d40-4706-8a0b-e95259e192ba-kube-api-access-dpcct" (OuterVolumeSpecName: "kube-api-access-dpcct") pod "866c1e6f-1d40-4706-8a0b-e95259e192ba" (UID: "866c1e6f-1d40-4706-8a0b-e95259e192ba"). InnerVolumeSpecName "kube-api-access-dpcct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.835685 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "866c1e6f-1d40-4706-8a0b-e95259e192ba" (UID: "866c1e6f-1d40-4706-8a0b-e95259e192ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.869953 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.869997 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpcct\" (UniqueName: \"kubernetes.io/projected/866c1e6f-1d40-4706-8a0b-e95259e192ba-kube-api-access-dpcct\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:23 crc kubenswrapper[4912]: I1203 01:48:23.870047 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/866c1e6f-1d40-4706-8a0b-e95259e192ba-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:24 crc kubenswrapper[4912]: I1203 01:48:24.503571 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cqmld" Dec 03 01:48:24 crc kubenswrapper[4912]: I1203 01:48:24.594843 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cqmld"] Dec 03 01:48:24 crc kubenswrapper[4912]: I1203 01:48:24.594922 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cqmld"] Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.169435 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.340679 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceph\") pod \"e34cc78e-e946-4d45-a3da-962e308ad972\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.340759 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-1\") pod \"e34cc78e-e946-4d45-a3da-962e308ad972\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.340818 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-0\") pod \"e34cc78e-e946-4d45-a3da-962e308ad972\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.340893 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ssh-key\") pod \"e34cc78e-e946-4d45-a3da-962e308ad972\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.340939 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-inventory\") pod \"e34cc78e-e946-4d45-a3da-962e308ad972\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.340980 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-telemetry-power-monitoring-combined-ca-bundle\") pod \"e34cc78e-e946-4d45-a3da-962e308ad972\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.341073 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rxnj\" (UniqueName: \"kubernetes.io/projected/e34cc78e-e946-4d45-a3da-962e308ad972-kube-api-access-4rxnj\") pod \"e34cc78e-e946-4d45-a3da-962e308ad972\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.341184 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-2\") pod \"e34cc78e-e946-4d45-a3da-962e308ad972\" (UID: \"e34cc78e-e946-4d45-a3da-962e308ad972\") " Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.347764 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "e34cc78e-e946-4d45-a3da-962e308ad972" (UID: "e34cc78e-e946-4d45-a3da-962e308ad972"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.349149 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceph" (OuterVolumeSpecName: "ceph") pod "e34cc78e-e946-4d45-a3da-962e308ad972" (UID: "e34cc78e-e946-4d45-a3da-962e308ad972"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.350779 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e34cc78e-e946-4d45-a3da-962e308ad972-kube-api-access-4rxnj" (OuterVolumeSpecName: "kube-api-access-4rxnj") pod "e34cc78e-e946-4d45-a3da-962e308ad972" (UID: "e34cc78e-e946-4d45-a3da-962e308ad972"). InnerVolumeSpecName "kube-api-access-4rxnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.375556 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e34cc78e-e946-4d45-a3da-962e308ad972" (UID: "e34cc78e-e946-4d45-a3da-962e308ad972"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.379198 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-inventory" (OuterVolumeSpecName: "inventory") pod "e34cc78e-e946-4d45-a3da-962e308ad972" (UID: "e34cc78e-e946-4d45-a3da-962e308ad972"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.381953 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "e34cc78e-e946-4d45-a3da-962e308ad972" (UID: "e34cc78e-e946-4d45-a3da-962e308ad972"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.383976 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "e34cc78e-e946-4d45-a3da-962e308ad972" (UID: "e34cc78e-e946-4d45-a3da-962e308ad972"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.388337 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "e34cc78e-e946-4d45-a3da-962e308ad972" (UID: "e34cc78e-e946-4d45-a3da-962e308ad972"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.444919 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rxnj\" (UniqueName: \"kubernetes.io/projected/e34cc78e-e946-4d45-a3da-962e308ad972-kube-api-access-4rxnj\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.445151 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.445270 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.445352 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.445968 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.446153 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.446257 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.446572 4912 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e34cc78e-e946-4d45-a3da-962e308ad972-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.517633 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" event={"ID":"e34cc78e-e946-4d45-a3da-962e308ad972","Type":"ContainerDied","Data":"6193ce126c62ff3b76461baafcdc616bf4d6186a19c6bf34efdd9feb287ec8c9"} Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.517678 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6193ce126c62ff3b76461baafcdc616bf4d6186a19c6bf34efdd9feb287ec8c9" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.517725 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.610414 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t"] Dec 03 01:48:25 crc kubenswrapper[4912]: E1203 01:48:25.612050 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerName="registry-server" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.612130 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerName="registry-server" Dec 03 01:48:25 crc kubenswrapper[4912]: E1203 01:48:25.612226 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerName="extract-utilities" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.612291 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerName="extract-utilities" Dec 03 01:48:25 crc kubenswrapper[4912]: E1203 01:48:25.612367 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34cc78e-e946-4d45-a3da-962e308ad972" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.612454 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34cc78e-e946-4d45-a3da-962e308ad972" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 01:48:25 crc kubenswrapper[4912]: E1203 01:48:25.612546 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerName="extract-content" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.612612 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerName="extract-content" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.612930 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="866c1e6f-1d40-4706-8a0b-e95259e192ba" containerName="registry-server" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.613024 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="e34cc78e-e946-4d45-a3da-962e308ad972" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.614403 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.616284 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.616337 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.616815 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.616934 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-6gznk" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.618337 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.621246 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.628697 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t"] Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.650032 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.650184 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.650255 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr562\" (UniqueName: \"kubernetes.io/projected/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-kube-api-access-qr562\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.650300 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.650331 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.650370 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.752356 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.752455 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr562\" (UniqueName: \"kubernetes.io/projected/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-kube-api-access-qr562\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.752505 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.752527 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.752559 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.752647 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.756695 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.757040 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.757165 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.761149 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.762312 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.773619 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr562\" (UniqueName: \"kubernetes.io/projected/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-kube-api-access-qr562\") pod \"logging-edpm-deployment-openstack-edpm-ipam-46v4t\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:25 crc kubenswrapper[4912]: I1203 01:48:25.939918 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:26 crc kubenswrapper[4912]: I1203 01:48:26.596527 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="866c1e6f-1d40-4706-8a0b-e95259e192ba" path="/var/lib/kubelet/pods/866c1e6f-1d40-4706-8a0b-e95259e192ba/volumes" Dec 03 01:48:26 crc kubenswrapper[4912]: I1203 01:48:26.598222 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t"] Dec 03 01:48:27 crc kubenswrapper[4912]: I1203 01:48:27.546952 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" event={"ID":"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f","Type":"ContainerStarted","Data":"93e24bb9fe6e5e26989db8142da40d9fa12ac6de769b43fb66bef299ab9cf0dd"} Dec 03 01:48:27 crc kubenswrapper[4912]: I1203 01:48:27.547215 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" event={"ID":"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f","Type":"ContainerStarted","Data":"24eabe75f81c4a4a144d815422c644d1c25ba625eb5bb9545b27aa0327028308"} Dec 03 01:48:27 crc kubenswrapper[4912]: I1203 01:48:27.578302 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" podStartSLOduration=2.074517444 podStartE2EDuration="2.578285738s" podCreationTimestamp="2025-12-03 01:48:25 +0000 UTC" firstStartedPulling="2025-12-03 01:48:26.57780901 +0000 UTC m=+5092.219829570" lastFinishedPulling="2025-12-03 01:48:27.081577264 +0000 UTC m=+5092.723597864" observedRunningTime="2025-12-03 01:48:27.572396759 +0000 UTC m=+5093.214417319" watchObservedRunningTime="2025-12-03 01:48:27.578285738 +0000 UTC m=+5093.220306298" Dec 03 01:48:37 crc kubenswrapper[4912]: I1203 01:48:37.572247 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:48:37 crc kubenswrapper[4912]: E1203 01:48:37.573301 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:48:43 crc kubenswrapper[4912]: I1203 01:48:43.761690 4912 generic.go:334] "Generic (PLEG): container finished" podID="77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" containerID="93e24bb9fe6e5e26989db8142da40d9fa12ac6de769b43fb66bef299ab9cf0dd" exitCode=0 Dec 03 01:48:43 crc kubenswrapper[4912]: I1203 01:48:43.761996 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" event={"ID":"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f","Type":"ContainerDied","Data":"93e24bb9fe6e5e26989db8142da40d9fa12ac6de769b43fb66bef299ab9cf0dd"} Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.272401 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.371743 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr562\" (UniqueName: \"kubernetes.io/projected/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-kube-api-access-qr562\") pod \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.371838 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-0\") pod \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.371879 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ssh-key\") pod \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.371992 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ceph\") pod \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.372140 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-1\") pod \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.372198 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-inventory\") pod \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\" (UID: \"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f\") " Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.379244 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ceph" (OuterVolumeSpecName: "ceph") pod "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" (UID: "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.380275 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-kube-api-access-qr562" (OuterVolumeSpecName: "kube-api-access-qr562") pod "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" (UID: "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f"). InnerVolumeSpecName "kube-api-access-qr562". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.419282 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" (UID: "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.475204 4912 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.475248 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr562\" (UniqueName: \"kubernetes.io/projected/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-kube-api-access-qr562\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.475261 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.791635 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" event={"ID":"77e171a1-53b6-4a5d-94a2-2d9b42e1e71f","Type":"ContainerDied","Data":"24eabe75f81c4a4a144d815422c644d1c25ba625eb5bb9545b27aa0327028308"} Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.791710 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24eabe75f81c4a4a144d815422c644d1c25ba625eb5bb9545b27aa0327028308" Dec 03 01:48:45 crc kubenswrapper[4912]: I1203 01:48:45.791734 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-46v4t" Dec 03 01:48:46 crc kubenswrapper[4912]: I1203 01:48:46.326293 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-inventory" (OuterVolumeSpecName: "inventory") pod "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" (UID: "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:46 crc kubenswrapper[4912]: I1203 01:48:46.332341 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" (UID: "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:46 crc kubenswrapper[4912]: I1203 01:48:46.335155 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" (UID: "77e171a1-53b6-4a5d-94a2-2d9b42e1e71f"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:48:46 crc kubenswrapper[4912]: I1203 01:48:46.412562 4912 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:46 crc kubenswrapper[4912]: I1203 01:48:46.412597 4912 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:46 crc kubenswrapper[4912]: I1203 01:48:46.412611 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/77e171a1-53b6-4a5d-94a2-2d9b42e1e71f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:48:51 crc kubenswrapper[4912]: I1203 01:48:51.571953 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:48:51 crc kubenswrapper[4912]: E1203 01:48:51.572617 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.136602 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s5kr2"] Dec 03 01:48:59 crc kubenswrapper[4912]: E1203 01:48:59.138474 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.138516 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.139199 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="77e171a1-53b6-4a5d-94a2-2d9b42e1e71f" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.142784 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.145811 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s5kr2"] Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.265501 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-catalog-content\") pod \"community-operators-s5kr2\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.265574 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv8mh\" (UniqueName: \"kubernetes.io/projected/96532d70-09e3-4830-a4e7-1349fb2e602a-kube-api-access-nv8mh\") pod \"community-operators-s5kr2\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.265653 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-utilities\") pod \"community-operators-s5kr2\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.367828 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-catalog-content\") pod \"community-operators-s5kr2\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.367872 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv8mh\" (UniqueName: \"kubernetes.io/projected/96532d70-09e3-4830-a4e7-1349fb2e602a-kube-api-access-nv8mh\") pod \"community-operators-s5kr2\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.367909 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-utilities\") pod \"community-operators-s5kr2\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.368319 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-catalog-content\") pod \"community-operators-s5kr2\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.368369 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-utilities\") pod \"community-operators-s5kr2\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.502727 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv8mh\" (UniqueName: \"kubernetes.io/projected/96532d70-09e3-4830-a4e7-1349fb2e602a-kube-api-access-nv8mh\") pod \"community-operators-s5kr2\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:48:59 crc kubenswrapper[4912]: I1203 01:48:59.779843 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:49:00 crc kubenswrapper[4912]: I1203 01:49:00.331055 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s5kr2"] Dec 03 01:49:00 crc kubenswrapper[4912]: I1203 01:49:00.985487 4912 generic.go:334] "Generic (PLEG): container finished" podID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerID="23b03489859f570820bd666c45f9d53175721d55978ca02df11f21c9c1ccc16f" exitCode=0 Dec 03 01:49:00 crc kubenswrapper[4912]: I1203 01:49:00.985634 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5kr2" event={"ID":"96532d70-09e3-4830-a4e7-1349fb2e602a","Type":"ContainerDied","Data":"23b03489859f570820bd666c45f9d53175721d55978ca02df11f21c9c1ccc16f"} Dec 03 01:49:00 crc kubenswrapper[4912]: I1203 01:49:00.985988 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5kr2" event={"ID":"96532d70-09e3-4830-a4e7-1349fb2e602a","Type":"ContainerStarted","Data":"c376e4ec8e13f8919ddee2d09beec5efaa12682017dd3a4646d21462b1b7874e"} Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.836034 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.837940 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.845300 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.846061 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.852723 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.924331 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-dev\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.924518 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.924589 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.924645 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-scripts\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.924719 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-ceph\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.924744 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8cfg\" (UniqueName: \"kubernetes.io/projected/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-kube-api-access-b8cfg\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.924812 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.924851 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-sys\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.924889 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.925009 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.925057 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.925101 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-config-data\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.925246 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.925322 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-lib-modules\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.925382 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.925409 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-run\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.930302 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.933991 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.941341 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 03 01:49:01 crc kubenswrapper[4912]: I1203 01:49:01.952394 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027012 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027104 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027140 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027177 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027219 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-run\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027241 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027397 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027458 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027495 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-config-data\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027688 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027746 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027763 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027784 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027810 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027837 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027862 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-lib-modules\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027898 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027920 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-run\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027935 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027981 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3e1c22b1-b578-4957-98b5-b981f1f7dee2-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.027998 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-dev\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028054 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028073 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028094 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028119 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-scripts\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028151 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028176 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8cfg\" (UniqueName: \"kubernetes.io/projected/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-kube-api-access-b8cfg\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028193 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-ceph\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028210 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028234 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028256 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5d7b\" (UniqueName: \"kubernetes.io/projected/3e1c22b1-b578-4957-98b5-b981f1f7dee2-kube-api-access-l5d7b\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028279 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-sys\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028301 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028325 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028573 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-lib-modules\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028596 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028632 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-sys\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028632 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028650 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-run\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028931 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-dev\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.028974 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.029223 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.033555 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.034095 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.034153 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-scripts\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.034618 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-ceph\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.034802 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-config-data\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.044093 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8cfg\" (UniqueName: \"kubernetes.io/projected/51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7-kube-api-access-b8cfg\") pod \"cinder-backup-0\" (UID: \"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7\") " pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129627 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129690 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129732 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129752 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5d7b\" (UniqueName: \"kubernetes.io/projected/3e1c22b1-b578-4957-98b5-b981f1f7dee2-kube-api-access-l5d7b\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129779 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129804 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129820 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129842 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129868 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-run\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129886 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129931 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129947 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129968 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129986 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130019 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130048 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3e1c22b1-b578-4957-98b5-b981f1f7dee2-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.129747 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130333 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130546 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130597 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-run\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130608 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130653 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-dev\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130654 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130652 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-sys\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130693 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.130700 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3e1c22b1-b578-4957-98b5-b981f1f7dee2-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.132744 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.133204 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.133326 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3e1c22b1-b578-4957-98b5-b981f1f7dee2-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.133508 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.139227 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e1c22b1-b578-4957-98b5-b981f1f7dee2-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.146056 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5d7b\" (UniqueName: \"kubernetes.io/projected/3e1c22b1-b578-4957-98b5-b981f1f7dee2-kube-api-access-l5d7b\") pod \"cinder-volume-volume1-0\" (UID: \"3e1c22b1-b578-4957-98b5-b981f1f7dee2\") " pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.216659 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.259884 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.724541 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.726695 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.731189 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.731671 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bvvth" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.731793 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.731985 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.734188 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.754769 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-jg4rj"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.756231 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.772914 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-d0c6-account-create-update-r2rnf"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.776129 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.780667 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.790790 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.799112 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.812768 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.813825 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.820675 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-jg4rj"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.837723 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-d0c6-account-create-update-r2rnf"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849099 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-logs\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849177 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69661ff8-038d-4f96-b806-877ec5aaa6ba-operator-scripts\") pod \"manila-d0c6-account-create-update-r2rnf\" (UID: \"69661ff8-038d-4f96-b806-877ec5aaa6ba\") " pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849209 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849262 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t24pj\" (UniqueName: \"kubernetes.io/projected/69661ff8-038d-4f96-b806-877ec5aaa6ba-kube-api-access-t24pj\") pod \"manila-d0c6-account-create-update-r2rnf\" (UID: \"69661ff8-038d-4f96-b806-877ec5aaa6ba\") " pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849304 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849341 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849391 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5kvl\" (UniqueName: \"kubernetes.io/projected/86aa696b-0e24-4cde-9944-4142d0fb81a8-kube-api-access-x5kvl\") pod \"manila-db-create-jg4rj\" (UID: \"86aa696b-0e24-4cde-9944-4142d0fb81a8\") " pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849558 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849617 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849669 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86aa696b-0e24-4cde-9944-4142d0fb81a8-operator-scripts\") pod \"manila-db-create-jg4rj\" (UID: \"86aa696b-0e24-4cde-9944-4142d0fb81a8\") " pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849712 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4hbl\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-kube-api-access-r4hbl\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849745 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.849772 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.863841 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.937553 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:02 crc kubenswrapper[4912]: E1203 01:49:02.938334 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run kube-api-access-7gwfk logs public-tls-certs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-external-api-0" podUID="5a27f7e0-a47d-46ca-b395-259b856d2c33" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952191 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-scripts\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952253 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952296 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952357 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86aa696b-0e24-4cde-9944-4142d0fb81a8-operator-scripts\") pod \"manila-db-create-jg4rj\" (UID: \"86aa696b-0e24-4cde-9944-4142d0fb81a8\") " pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952389 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-ceph\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952445 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952481 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4hbl\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-kube-api-access-r4hbl\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952516 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952547 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952590 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gwfk\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-kube-api-access-7gwfk\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952611 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952636 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-logs\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952670 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69661ff8-038d-4f96-b806-877ec5aaa6ba-operator-scripts\") pod \"manila-d0c6-account-create-update-r2rnf\" (UID: \"69661ff8-038d-4f96-b806-877ec5aaa6ba\") " pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952689 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952706 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952736 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-config-data\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952766 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t24pj\" (UniqueName: \"kubernetes.io/projected/69661ff8-038d-4f96-b806-877ec5aaa6ba-kube-api-access-t24pj\") pod \"manila-d0c6-account-create-update-r2rnf\" (UID: \"69661ff8-038d-4f96-b806-877ec5aaa6ba\") " pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952788 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952812 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952834 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-logs\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952853 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.952888 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5kvl\" (UniqueName: \"kubernetes.io/projected/86aa696b-0e24-4cde-9944-4142d0fb81a8-kube-api-access-x5kvl\") pod \"manila-db-create-jg4rj\" (UID: \"86aa696b-0e24-4cde-9944-4142d0fb81a8\") " pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.953841 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.957372 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.958170 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69661ff8-038d-4f96-b806-877ec5aaa6ba-operator-scripts\") pod \"manila-d0c6-account-create-update-r2rnf\" (UID: \"69661ff8-038d-4f96-b806-877ec5aaa6ba\") " pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.961645 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-logs\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.964128 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86aa696b-0e24-4cde-9944-4142d0fb81a8-operator-scripts\") pod \"manila-db-create-jg4rj\" (UID: \"86aa696b-0e24-4cde-9944-4142d0fb81a8\") " pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.965063 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-69f9d58fc9-6l6qj"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.966871 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.967776 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.977002 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.977146 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-87lm6" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.977144 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69f9d58fc9-6l6qj"] Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.977187 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.989296 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.989550 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 03 01:49:02 crc kubenswrapper[4912]: I1203 01:49:02.990545 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.018859 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5kvl\" (UniqueName: \"kubernetes.io/projected/86aa696b-0e24-4cde-9944-4142d0fb81a8-kube-api-access-x5kvl\") pod \"manila-db-create-jg4rj\" (UID: \"86aa696b-0e24-4cde-9944-4142d0fb81a8\") " pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.019352 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.029771 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4hbl\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-kube-api-access-r4hbl\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.030344 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-ceph\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.030776 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t24pj\" (UniqueName: \"kubernetes.io/projected/69661ff8-038d-4f96-b806-877ec5aaa6ba-kube-api-access-t24pj\") pod \"manila-d0c6-account-create-update-r2rnf\" (UID: \"69661ff8-038d-4f96-b806-877ec5aaa6ba\") " pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.061728 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-horizon-secret-key\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.063061 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gwfk\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-kube-api-access-7gwfk\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.063108 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.063140 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-scripts\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.063234 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.063287 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-config-data\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.063982 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.064020 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-logs\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.064124 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-scripts\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.064241 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-ceph\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.064267 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-logs\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.064307 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.064327 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-config-data\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.064396 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwv7g\" (UniqueName: \"kubernetes.io/projected/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-kube-api-access-bwv7g\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.065224 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-logs\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.065566 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.066227 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.078717 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.078791 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-config-data\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.085769 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.086993 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.093814 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-scripts\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.095682 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5kr2" event={"ID":"96532d70-09e3-4830-a4e7-1349fb2e602a","Type":"ContainerStarted","Data":"8c0022724bb5acd03361b37a668e5e0049cb7db1f2a821d9ae63dbdc13822123"} Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.103598 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.104105 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7","Type":"ContainerStarted","Data":"c29a26cf8d5b544175c31458762bd0e0f4da802d6b08d6a3920df48993252f0b"} Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.104494 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-ceph\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.105621 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.114642 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gwfk\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-kube-api-access-7gwfk\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.122705 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.136482 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.136571 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-77dc7f54bf-f2rhc"] Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.139548 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.141555 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.163168 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.167082 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-logs\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.167636 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-logs\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.168277 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-config-data\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.168337 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwv7g\" (UniqueName: \"kubernetes.io/projected/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-kube-api-access-bwv7g\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.168561 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-horizon-secret-key\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.168662 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-scripts\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.169404 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-config-data\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.171300 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-scripts\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.175292 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-horizon-secret-key\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.179158 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77dc7f54bf-f2rhc"] Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.188493 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwv7g\" (UniqueName: \"kubernetes.io/projected/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-kube-api-access-bwv7g\") pod \"horizon-69f9d58fc9-6l6qj\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.189551 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.237311 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.278352 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-logs\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.278747 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-horizon-secret-key\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.278775 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-config-data\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.279061 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4tbq\" (UniqueName: \"kubernetes.io/projected/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-kube-api-access-s4tbq\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.284256 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-scripts\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.316117 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.317140 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.389356 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"5a27f7e0-a47d-46ca-b395-259b856d2c33\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.389594 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-combined-ca-bundle\") pod \"5a27f7e0-a47d-46ca-b395-259b856d2c33\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.389614 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gwfk\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-kube-api-access-7gwfk\") pod \"5a27f7e0-a47d-46ca-b395-259b856d2c33\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.389758 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-ceph\") pod \"5a27f7e0-a47d-46ca-b395-259b856d2c33\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.389796 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-public-tls-certs\") pod \"5a27f7e0-a47d-46ca-b395-259b856d2c33\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.389844 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-httpd-run\") pod \"5a27f7e0-a47d-46ca-b395-259b856d2c33\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.389878 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-logs\") pod \"5a27f7e0-a47d-46ca-b395-259b856d2c33\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.394774 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5a27f7e0-a47d-46ca-b395-259b856d2c33" (UID: "5a27f7e0-a47d-46ca-b395-259b856d2c33"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.395324 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-logs" (OuterVolumeSpecName: "logs") pod "5a27f7e0-a47d-46ca-b395-259b856d2c33" (UID: "5a27f7e0-a47d-46ca-b395-259b856d2c33"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.395649 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-kube-api-access-7gwfk" (OuterVolumeSpecName: "kube-api-access-7gwfk") pod "5a27f7e0-a47d-46ca-b395-259b856d2c33" (UID: "5a27f7e0-a47d-46ca-b395-259b856d2c33"). InnerVolumeSpecName "kube-api-access-7gwfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.396162 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5a27f7e0-a47d-46ca-b395-259b856d2c33" (UID: "5a27f7e0-a47d-46ca-b395-259b856d2c33"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.396690 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "5a27f7e0-a47d-46ca-b395-259b856d2c33" (UID: "5a27f7e0-a47d-46ca-b395-259b856d2c33"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.399116 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a27f7e0-a47d-46ca-b395-259b856d2c33" (UID: "5a27f7e0-a47d-46ca-b395-259b856d2c33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.401471 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-ceph" (OuterVolumeSpecName: "ceph") pod "5a27f7e0-a47d-46ca-b395-259b856d2c33" (UID: "5a27f7e0-a47d-46ca-b395-259b856d2c33"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.403615 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-scripts\") pod \"5a27f7e0-a47d-46ca-b395-259b856d2c33\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.403688 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-config-data\") pod \"5a27f7e0-a47d-46ca-b395-259b856d2c33\" (UID: \"5a27f7e0-a47d-46ca-b395-259b856d2c33\") " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.404311 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4tbq\" (UniqueName: \"kubernetes.io/projected/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-kube-api-access-s4tbq\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.404377 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-scripts\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.404621 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-logs\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.404682 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-horizon-secret-key\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.405016 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-config-data\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.405872 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-scripts\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.407975 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gwfk\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-kube-api-access-7gwfk\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.408773 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-horizon-secret-key\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.409143 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.409158 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a27f7e0-a47d-46ca-b395-259b856d2c33-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.409167 4912 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.409176 4912 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.409184 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a27f7e0-a47d-46ca-b395-259b856d2c33-logs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.409187 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-logs\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.409212 4912 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.409300 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-config-data" (OuterVolumeSpecName: "config-data") pod "5a27f7e0-a47d-46ca-b395-259b856d2c33" (UID: "5a27f7e0-a47d-46ca-b395-259b856d2c33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.409657 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-config-data\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.412493 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-scripts" (OuterVolumeSpecName: "scripts") pod "5a27f7e0-a47d-46ca-b395-259b856d2c33" (UID: "5a27f7e0-a47d-46ca-b395-259b856d2c33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.430951 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4tbq\" (UniqueName: \"kubernetes.io/projected/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-kube-api-access-s4tbq\") pod \"horizon-77dc7f54bf-f2rhc\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.434980 4912 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.511042 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.511301 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a27f7e0-a47d-46ca-b395-259b856d2c33-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.511314 4912 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.624210 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-jg4rj"] Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.645946 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.890569 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-d0c6-account-create-update-r2rnf"] Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.907126 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-69f9d58fc9-6l6qj"] Dec 03 01:49:03 crc kubenswrapper[4912]: I1203 01:49:03.941184 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 01:49:04 crc kubenswrapper[4912]: W1203 01:49:04.000678 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f9ca12c_238d_47f3_a39d_6c2df17d6baa.slice/crio-cdc217ebc1f6e0d5324896094ae6191595a7c4edfa86038142ef7ed1069505d2 WatchSource:0}: Error finding container cdc217ebc1f6e0d5324896094ae6191595a7c4edfa86038142ef7ed1069505d2: Status 404 returned error can't find the container with id cdc217ebc1f6e0d5324896094ae6191595a7c4edfa86038142ef7ed1069505d2 Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.135832 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00","Type":"ContainerStarted","Data":"c81cfc7346f5eacda446e987fad10090ea2784fd6856ae40cb39f19bae7861a5"} Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.139720 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-jg4rj" event={"ID":"86aa696b-0e24-4cde-9944-4142d0fb81a8","Type":"ContainerStarted","Data":"4afcfb2b941810d05bb5ba3307bfe70475956503b8913190ff8e426fb6791268"} Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.139748 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-jg4rj" event={"ID":"86aa696b-0e24-4cde-9944-4142d0fb81a8","Type":"ContainerStarted","Data":"a493249dc4c3eb0474333f52cd13886fd540c6815cfa2aeee49428bbefe5fe87"} Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.142855 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-d0c6-account-create-update-r2rnf" event={"ID":"69661ff8-038d-4f96-b806-877ec5aaa6ba","Type":"ContainerStarted","Data":"0bf21df4b73a2f313f3823ab907e1556f1dc49aed382eb16b656f4bde04faf38"} Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.144278 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69f9d58fc9-6l6qj" event={"ID":"7f9ca12c-238d-47f3-a39d-6c2df17d6baa","Type":"ContainerStarted","Data":"cdc217ebc1f6e0d5324896094ae6191595a7c4edfa86038142ef7ed1069505d2"} Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.145308 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3e1c22b1-b578-4957-98b5-b981f1f7dee2","Type":"ContainerStarted","Data":"4a442b0a47c854037d376d85b7a124fc115df969c6a5afe4e7f6801085dc9ba4"} Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.154448 4912 generic.go:334] "Generic (PLEG): container finished" podID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerID="8c0022724bb5acd03361b37a668e5e0049cb7db1f2a821d9ae63dbdc13822123" exitCode=0 Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.154680 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.154672 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5kr2" event={"ID":"96532d70-09e3-4830-a4e7-1349fb2e602a","Type":"ContainerDied","Data":"8c0022724bb5acd03361b37a668e5e0049cb7db1f2a821d9ae63dbdc13822123"} Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.157509 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-create-jg4rj" podStartSLOduration=2.15749362 podStartE2EDuration="2.15749362s" podCreationTimestamp="2025-12-03 01:49:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:49:04.154578152 +0000 UTC m=+5129.796598722" watchObservedRunningTime="2025-12-03 01:49:04.15749362 +0000 UTC m=+5129.799514180" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.219590 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77dc7f54bf-f2rhc"] Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.329971 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.345355 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.355780 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.366834 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.371508 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.375605 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.398353 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.450644 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.451134 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.451167 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.451198 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-ceph\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.451252 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-config-data\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.451355 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-scripts\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.451395 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-logs\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.451420 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms55t\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-kube-api-access-ms55t\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.451464 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569126 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-ceph\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569252 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-config-data\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569325 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-scripts\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569364 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-logs\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569397 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms55t\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-kube-api-access-ms55t\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569497 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569535 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569557 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569585 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.569887 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.572793 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-logs\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.573263 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.588624 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-config-data\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.588666 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-ceph\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.589023 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.593739 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-scripts\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.596318 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms55t\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-kube-api-access-ms55t\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.596795 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.602178 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a27f7e0-a47d-46ca-b395-259b856d2c33" path="/var/lib/kubelet/pods/5a27f7e0-a47d-46ca-b395-259b856d2c33/volumes" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.674131 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:04 crc kubenswrapper[4912]: I1203 01:49:04.720785 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.181528 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77dc7f54bf-f2rhc" event={"ID":"a6880b8c-c092-43bd-9cd1-8c837be9bfe2","Type":"ContainerStarted","Data":"4f2329ffd101326d55929ee87f1a21abb689adb84d4e39785b6ea1ff6cfae440"} Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.183942 4912 generic.go:334] "Generic (PLEG): container finished" podID="86aa696b-0e24-4cde-9944-4142d0fb81a8" containerID="4afcfb2b941810d05bb5ba3307bfe70475956503b8913190ff8e426fb6791268" exitCode=0 Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.184520 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-jg4rj" event={"ID":"86aa696b-0e24-4cde-9944-4142d0fb81a8","Type":"ContainerDied","Data":"4afcfb2b941810d05bb5ba3307bfe70475956503b8913190ff8e426fb6791268"} Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.190856 4912 generic.go:334] "Generic (PLEG): container finished" podID="69661ff8-038d-4f96-b806-877ec5aaa6ba" containerID="783a6eb0f0944208070d0e047330fdb365a5e286bcb47bb82a99a5e0f047230b" exitCode=0 Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.190915 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-d0c6-account-create-update-r2rnf" event={"ID":"69661ff8-038d-4f96-b806-877ec5aaa6ba","Type":"ContainerDied","Data":"783a6eb0f0944208070d0e047330fdb365a5e286bcb47bb82a99a5e0f047230b"} Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.203527 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7","Type":"ContainerStarted","Data":"a7926611660d8086686912d6b71fccc1315c2018aa45b704576941e9270cbda1"} Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.206054 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3e1c22b1-b578-4957-98b5-b981f1f7dee2","Type":"ContainerStarted","Data":"5faefdf63380fc63e10821173c77ccb1cc11582662ac37b33681ea75a7e1424f"} Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.515904 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:05 crc kubenswrapper[4912]: W1203 01:49:05.517025 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod700218d5_259f_478b_94b3_abc87dc8c76f.slice/crio-7ad369b588d7ebf1f91c99647daf02911bfd09c96fb089ebae91399582fc9aab WatchSource:0}: Error finding container 7ad369b588d7ebf1f91c99647daf02911bfd09c96fb089ebae91399582fc9aab: Status 404 returned error can't find the container with id 7ad369b588d7ebf1f91c99647daf02911bfd09c96fb089ebae91399582fc9aab Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.571790 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:49:05 crc kubenswrapper[4912]: E1203 01:49:05.572029 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.944373 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69f9d58fc9-6l6qj"] Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.977084 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:05 crc kubenswrapper[4912]: I1203 01:49:05.998392 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d9dc8d44b-7tljg"] Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.000372 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.005218 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.015690 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d9dc8d44b-7tljg"] Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.050726 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-config-data\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.050796 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08762529-e1c5-4358-bc43-c0d81a4c6df6-logs\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.050838 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-tls-certs\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.050858 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-combined-ca-bundle\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.050895 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-scripts\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.050951 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-secret-key\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.050987 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcxrz\" (UniqueName: \"kubernetes.io/projected/08762529-e1c5-4358-bc43-c0d81a4c6df6-kube-api-access-tcxrz\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.082936 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77dc7f54bf-f2rhc"] Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.151460 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d4dbbf8d-5vgmn"] Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.153451 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.153857 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-config-data\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.153925 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08762529-e1c5-4358-bc43-c0d81a4c6df6-logs\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.153952 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-tls-certs\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.153972 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-combined-ca-bundle\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.154020 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-scripts\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.154078 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-secret-key\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.154114 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcxrz\" (UniqueName: \"kubernetes.io/projected/08762529-e1c5-4358-bc43-c0d81a4c6df6-kube-api-access-tcxrz\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.155716 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08762529-e1c5-4358-bc43-c0d81a4c6df6-logs\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.156201 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-scripts\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.156647 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-config-data\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.161990 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-tls-certs\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.177358 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-combined-ca-bundle\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.182098 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-secret-key\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.182547 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcxrz\" (UniqueName: \"kubernetes.io/projected/08762529-e1c5-4358-bc43-c0d81a4c6df6-kube-api-access-tcxrz\") pod \"horizon-d9dc8d44b-7tljg\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.193784 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d4dbbf8d-5vgmn"] Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.232313 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5kr2" event={"ID":"96532d70-09e3-4830-a4e7-1349fb2e602a","Type":"ContainerStarted","Data":"4c1ebe643ab6278e030ac9d71815b994375f9bb5d498c8c351f80604b0deec72"} Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.239217 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"700218d5-259f-478b-94b3-abc87dc8c76f","Type":"ContainerStarted","Data":"7ad369b588d7ebf1f91c99647daf02911bfd09c96fb089ebae91399582fc9aab"} Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.247911 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00","Type":"ContainerStarted","Data":"cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c"} Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.250081 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7","Type":"ContainerStarted","Data":"587086d12a48cfd16598177b2165bcc1b73ea1ebacce1a1c057b503ac8345f8c"} Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.256019 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d92b6a2-0079-4d39-9c11-9a31c22429f4-horizon-secret-key\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.256083 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d92b6a2-0079-4d39-9c11-9a31c22429f4-combined-ca-bundle\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.256130 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ndfd\" (UniqueName: \"kubernetes.io/projected/3d92b6a2-0079-4d39-9c11-9a31c22429f4-kube-api-access-6ndfd\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.256167 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d92b6a2-0079-4d39-9c11-9a31c22429f4-scripts\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.256266 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d92b6a2-0079-4d39-9c11-9a31c22429f4-horizon-tls-certs\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.256321 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d92b6a2-0079-4d39-9c11-9a31c22429f4-logs\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.256338 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d92b6a2-0079-4d39-9c11-9a31c22429f4-config-data\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.263279 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s5kr2" podStartSLOduration=3.274238965 podStartE2EDuration="7.263259902s" podCreationTimestamp="2025-12-03 01:48:59 +0000 UTC" firstStartedPulling="2025-12-03 01:49:00.988646154 +0000 UTC m=+5126.630666754" lastFinishedPulling="2025-12-03 01:49:04.977667131 +0000 UTC m=+5130.619687691" observedRunningTime="2025-12-03 01:49:06.257561098 +0000 UTC m=+5131.899581668" watchObservedRunningTime="2025-12-03 01:49:06.263259902 +0000 UTC m=+5131.905280462" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.280515 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"3e1c22b1-b578-4957-98b5-b981f1f7dee2","Type":"ContainerStarted","Data":"3b4efdf5559cc5746f29a16dbaf5ba5f909e317f9fe9d0d4673ed5273d8833ee"} Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.322919 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=4.0526303 podStartE2EDuration="5.322899239s" podCreationTimestamp="2025-12-03 01:49:01 +0000 UTC" firstStartedPulling="2025-12-03 01:49:03.017704248 +0000 UTC m=+5128.659724808" lastFinishedPulling="2025-12-03 01:49:04.287973187 +0000 UTC m=+5129.929993747" observedRunningTime="2025-12-03 01:49:06.313635879 +0000 UTC m=+5131.955656439" watchObservedRunningTime="2025-12-03 01:49:06.322899239 +0000 UTC m=+5131.964919789" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.335867 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.202694434 podStartE2EDuration="5.335852838s" podCreationTimestamp="2025-12-03 01:49:01 +0000 UTC" firstStartedPulling="2025-12-03 01:49:03.185182871 +0000 UTC m=+5128.827203431" lastFinishedPulling="2025-12-03 01:49:04.318341285 +0000 UTC m=+5129.960361835" observedRunningTime="2025-12-03 01:49:06.332253981 +0000 UTC m=+5131.974274541" watchObservedRunningTime="2025-12-03 01:49:06.335852838 +0000 UTC m=+5131.977873398" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.338775 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.359105 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d92b6a2-0079-4d39-9c11-9a31c22429f4-logs\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.359151 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d92b6a2-0079-4d39-9c11-9a31c22429f4-config-data\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.359194 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d92b6a2-0079-4d39-9c11-9a31c22429f4-horizon-secret-key\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.359420 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d92b6a2-0079-4d39-9c11-9a31c22429f4-combined-ca-bundle\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.359503 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ndfd\" (UniqueName: \"kubernetes.io/projected/3d92b6a2-0079-4d39-9c11-9a31c22429f4-kube-api-access-6ndfd\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.359567 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d92b6a2-0079-4d39-9c11-9a31c22429f4-scripts\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.359704 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d92b6a2-0079-4d39-9c11-9a31c22429f4-horizon-tls-certs\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.359912 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d92b6a2-0079-4d39-9c11-9a31c22429f4-logs\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.361057 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d92b6a2-0079-4d39-9c11-9a31c22429f4-config-data\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.361563 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d92b6a2-0079-4d39-9c11-9a31c22429f4-scripts\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.366417 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d92b6a2-0079-4d39-9c11-9a31c22429f4-combined-ca-bundle\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.368049 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3d92b6a2-0079-4d39-9c11-9a31c22429f4-horizon-secret-key\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.370637 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d92b6a2-0079-4d39-9c11-9a31c22429f4-horizon-tls-certs\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.388202 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ndfd\" (UniqueName: \"kubernetes.io/projected/3d92b6a2-0079-4d39-9c11-9a31c22429f4-kube-api-access-6ndfd\") pod \"horizon-6d4dbbf8d-5vgmn\" (UID: \"3d92b6a2-0079-4d39-9c11-9a31c22429f4\") " pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.603200 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.851997 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.877346 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69661ff8-038d-4f96-b806-877ec5aaa6ba-operator-scripts\") pod \"69661ff8-038d-4f96-b806-877ec5aaa6ba\" (UID: \"69661ff8-038d-4f96-b806-877ec5aaa6ba\") " Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.877718 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t24pj\" (UniqueName: \"kubernetes.io/projected/69661ff8-038d-4f96-b806-877ec5aaa6ba-kube-api-access-t24pj\") pod \"69661ff8-038d-4f96-b806-877ec5aaa6ba\" (UID: \"69661ff8-038d-4f96-b806-877ec5aaa6ba\") " Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.880899 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69661ff8-038d-4f96-b806-877ec5aaa6ba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "69661ff8-038d-4f96-b806-877ec5aaa6ba" (UID: "69661ff8-038d-4f96-b806-877ec5aaa6ba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.890062 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69661ff8-038d-4f96-b806-877ec5aaa6ba-kube-api-access-t24pj" (OuterVolumeSpecName: "kube-api-access-t24pj") pod "69661ff8-038d-4f96-b806-877ec5aaa6ba" (UID: "69661ff8-038d-4f96-b806-877ec5aaa6ba"). InnerVolumeSpecName "kube-api-access-t24pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.980984 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/69661ff8-038d-4f96-b806-877ec5aaa6ba-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:06 crc kubenswrapper[4912]: I1203 01:49:06.981016 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t24pj\" (UniqueName: \"kubernetes.io/projected/69661ff8-038d-4f96-b806-877ec5aaa6ba-kube-api-access-t24pj\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.099688 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.196576 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5kvl\" (UniqueName: \"kubernetes.io/projected/86aa696b-0e24-4cde-9944-4142d0fb81a8-kube-api-access-x5kvl\") pod \"86aa696b-0e24-4cde-9944-4142d0fb81a8\" (UID: \"86aa696b-0e24-4cde-9944-4142d0fb81a8\") " Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.196922 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86aa696b-0e24-4cde-9944-4142d0fb81a8-operator-scripts\") pod \"86aa696b-0e24-4cde-9944-4142d0fb81a8\" (UID: \"86aa696b-0e24-4cde-9944-4142d0fb81a8\") " Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.197996 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86aa696b-0e24-4cde-9944-4142d0fb81a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "86aa696b-0e24-4cde-9944-4142d0fb81a8" (UID: "86aa696b-0e24-4cde-9944-4142d0fb81a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.217705 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.252351 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86aa696b-0e24-4cde-9944-4142d0fb81a8-kube-api-access-x5kvl" (OuterVolumeSpecName: "kube-api-access-x5kvl") pod "86aa696b-0e24-4cde-9944-4142d0fb81a8" (UID: "86aa696b-0e24-4cde-9944-4142d0fb81a8"). InnerVolumeSpecName "kube-api-access-x5kvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.266424 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.296195 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"700218d5-259f-478b-94b3-abc87dc8c76f","Type":"ContainerStarted","Data":"4652d5b0bc1bd5458d9f5c40ead842bf1686b33f653f370fc6d8d50c4b4f8b65"} Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.299776 4912 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86aa696b-0e24-4cde-9944-4142d0fb81a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.299898 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5kvl\" (UniqueName: \"kubernetes.io/projected/86aa696b-0e24-4cde-9944-4142d0fb81a8-kube-api-access-x5kvl\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.300398 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00","Type":"ContainerStarted","Data":"fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179"} Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.300740 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerName="glance-log" containerID="cri-o://cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c" gracePeriod=30 Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.301223 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerName="glance-httpd" containerID="cri-o://fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179" gracePeriod=30 Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.312830 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-jg4rj" event={"ID":"86aa696b-0e24-4cde-9944-4142d0fb81a8","Type":"ContainerDied","Data":"a493249dc4c3eb0474333f52cd13886fd540c6815cfa2aeee49428bbefe5fe87"} Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.312864 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a493249dc4c3eb0474333f52cd13886fd540c6815cfa2aeee49428bbefe5fe87" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.312909 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-jg4rj" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.321390 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d9dc8d44b-7tljg"] Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.326335 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.326319517 podStartE2EDuration="6.326319517s" podCreationTimestamp="2025-12-03 01:49:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:49:07.32532449 +0000 UTC m=+5132.967345060" watchObservedRunningTime="2025-12-03 01:49:07.326319517 +0000 UTC m=+5132.968340077" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.334372 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-d0c6-account-create-update-r2rnf" event={"ID":"69661ff8-038d-4f96-b806-877ec5aaa6ba","Type":"ContainerDied","Data":"0bf21df4b73a2f313f3823ab907e1556f1dc49aed382eb16b656f4bde04faf38"} Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.334427 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bf21df4b73a2f313f3823ab907e1556f1dc49aed382eb16b656f4bde04faf38" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.334535 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-d0c6-account-create-update-r2rnf" Dec 03 01:49:07 crc kubenswrapper[4912]: I1203 01:49:07.482797 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d4dbbf8d-5vgmn"] Dec 03 01:49:07 crc kubenswrapper[4912]: W1203 01:49:07.491777 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d92b6a2_0079_4d39_9c11_9a31c22429f4.slice/crio-4059f2caa1190dc8cc4d21dd908ed21ca9cadc14f50a2bbcd4beb93aceafa4e4 WatchSource:0}: Error finding container 4059f2caa1190dc8cc4d21dd908ed21ca9cadc14f50a2bbcd4beb93aceafa4e4: Status 404 returned error can't find the container with id 4059f2caa1190dc8cc4d21dd908ed21ca9cadc14f50a2bbcd4beb93aceafa4e4 Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.252614 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.329276 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-config-data\") pod \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.329344 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-logs\") pod \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.329389 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-scripts\") pod \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.329423 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-ceph\") pod \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.329465 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-httpd-run\") pod \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.329515 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4hbl\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-kube-api-access-r4hbl\") pod \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.329551 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-combined-ca-bundle\") pod \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.329587 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-internal-tls-certs\") pod \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.329624 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\" (UID: \"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00\") " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.331106 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-logs" (OuterVolumeSpecName: "logs") pod "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" (UID: "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.340224 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" (UID: "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.341959 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-kube-api-access-r4hbl" (OuterVolumeSpecName: "kube-api-access-r4hbl") pod "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" (UID: "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00"). InnerVolumeSpecName "kube-api-access-r4hbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.344191 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-scripts" (OuterVolumeSpecName: "scripts") pod "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" (UID: "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.356243 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-h7pdq"] Dec 03 01:49:08 crc kubenswrapper[4912]: E1203 01:49:08.356748 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerName="glance-log" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.356761 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerName="glance-log" Dec 03 01:49:08 crc kubenswrapper[4912]: E1203 01:49:08.356780 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69661ff8-038d-4f96-b806-877ec5aaa6ba" containerName="mariadb-account-create-update" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.356786 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="69661ff8-038d-4f96-b806-877ec5aaa6ba" containerName="mariadb-account-create-update" Dec 03 01:49:08 crc kubenswrapper[4912]: E1203 01:49:08.356804 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86aa696b-0e24-4cde-9944-4142d0fb81a8" containerName="mariadb-database-create" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.356809 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="86aa696b-0e24-4cde-9944-4142d0fb81a8" containerName="mariadb-database-create" Dec 03 01:49:08 crc kubenswrapper[4912]: E1203 01:49:08.356821 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerName="glance-httpd" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.356826 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerName="glance-httpd" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.357017 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerName="glance-log" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.357028 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="69661ff8-038d-4f96-b806-877ec5aaa6ba" containerName="mariadb-account-create-update" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.357051 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerName="glance-httpd" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.357061 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="86aa696b-0e24-4cde-9944-4142d0fb81a8" containerName="mariadb-database-create" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.357508 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-ceph" (OuterVolumeSpecName: "ceph") pod "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" (UID: "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.359067 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.361929 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.362105 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-4ttg5" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.364993 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" (UID: "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.371968 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-h7pdq"] Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.403373 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4dbbf8d-5vgmn" event={"ID":"3d92b6a2-0079-4d39-9c11-9a31c22429f4","Type":"ContainerStarted","Data":"4059f2caa1190dc8cc4d21dd908ed21ca9cadc14f50a2bbcd4beb93aceafa4e4"} Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.408392 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" (UID: "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.414761 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"700218d5-259f-478b-94b3-abc87dc8c76f","Type":"ContainerStarted","Data":"76aefcb003ed1a03e9b7d4b58a3e9bdf63c9cec2ddcb6b5f9f90d8be255a0542"} Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.414900 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="700218d5-259f-478b-94b3-abc87dc8c76f" containerName="glance-log" containerID="cri-o://4652d5b0bc1bd5458d9f5c40ead842bf1686b33f653f370fc6d8d50c4b4f8b65" gracePeriod=30 Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.415306 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="700218d5-259f-478b-94b3-abc87dc8c76f" containerName="glance-httpd" containerID="cri-o://76aefcb003ed1a03e9b7d4b58a3e9bdf63c9cec2ddcb6b5f9f90d8be255a0542" gracePeriod=30 Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.422229 4912 generic.go:334] "Generic (PLEG): container finished" podID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerID="fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179" exitCode=0 Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.422261 4912 generic.go:334] "Generic (PLEG): container finished" podID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" containerID="cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c" exitCode=143 Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.422321 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00","Type":"ContainerDied","Data":"fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179"} Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.422332 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.422356 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00","Type":"ContainerDied","Data":"cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c"} Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.422370 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0846adb6-e9b8-4d98-ac3d-57fbf9d23a00","Type":"ContainerDied","Data":"c81cfc7346f5eacda446e987fad10090ea2784fd6856ae40cb39f19bae7861a5"} Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.422388 4912 scope.go:117] "RemoveContainer" containerID="fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.433251 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d9dc8d44b-7tljg" event={"ID":"08762529-e1c5-4358-bc43-c0d81a4c6df6","Type":"ContainerStarted","Data":"7a1396ea6289d0dbfc574b976e06d26a605d4e6fefca772b5a1356af744523f2"} Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435536 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-combined-ca-bundle\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435668 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-job-config-data\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435739 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wglrw\" (UniqueName: \"kubernetes.io/projected/fe7da98f-555b-47d1-ae2a-849d69e7540a-kube-api-access-wglrw\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435768 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-config-data\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435903 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4hbl\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-kube-api-access-r4hbl\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435925 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435949 4912 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435961 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-logs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435971 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435982 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.435992 4912 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.465732 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.465712358 podStartE2EDuration="4.465712358s" podCreationTimestamp="2025-12-03 01:49:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:49:08.439957365 +0000 UTC m=+5134.081977925" watchObservedRunningTime="2025-12-03 01:49:08.465712358 +0000 UTC m=+5134.107732928" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.476890 4912 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.489738 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-config-data" (OuterVolumeSpecName: "config-data") pod "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" (UID: "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.499300 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" (UID: "0846adb6-e9b8-4d98-ac3d-57fbf9d23a00"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.538626 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-combined-ca-bundle\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.539155 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-job-config-data\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.539517 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wglrw\" (UniqueName: \"kubernetes.io/projected/fe7da98f-555b-47d1-ae2a-849d69e7540a-kube-api-access-wglrw\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.539611 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-config-data\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.541282 4912 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.542550 4912 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.542648 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.546662 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-combined-ca-bundle\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.547211 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-job-config-data\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.548077 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-config-data\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.560124 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wglrw\" (UniqueName: \"kubernetes.io/projected/fe7da98f-555b-47d1-ae2a-849d69e7540a-kube-api-access-wglrw\") pod \"manila-db-sync-h7pdq\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.593185 4912 scope.go:117] "RemoveContainer" containerID="cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.764772 4912 scope.go:117] "RemoveContainer" containerID="fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179" Dec 03 01:49:08 crc kubenswrapper[4912]: E1203 01:49:08.766091 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179\": container with ID starting with fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179 not found: ID does not exist" containerID="fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.766152 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179"} err="failed to get container status \"fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179\": rpc error: code = NotFound desc = could not find container \"fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179\": container with ID starting with fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179 not found: ID does not exist" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.766178 4912 scope.go:117] "RemoveContainer" containerID="cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c" Dec 03 01:49:08 crc kubenswrapper[4912]: E1203 01:49:08.767278 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c\": container with ID starting with cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c not found: ID does not exist" containerID="cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.767320 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c"} err="failed to get container status \"cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c\": rpc error: code = NotFound desc = could not find container \"cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c\": container with ID starting with cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c not found: ID does not exist" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.770935 4912 scope.go:117] "RemoveContainer" containerID="fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.774523 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179"} err="failed to get container status \"fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179\": rpc error: code = NotFound desc = could not find container \"fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179\": container with ID starting with fce1ea72016a738c2d0ec32eb93d1f751ceb56c819b94131626389aebc47e179 not found: ID does not exist" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.774558 4912 scope.go:117] "RemoveContainer" containerID="cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.775139 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c"} err="failed to get container status \"cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c\": rpc error: code = NotFound desc = could not find container \"cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c\": container with ID starting with cdd8f95624b97d24b8d2f7544de3930ec1e9a97147984fd43dd245e88681889c not found: ID does not exist" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.790501 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.848292 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.851327 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.862071 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.864422 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.867518 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.869665 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.874743 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.994745 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17260561-5216-45fa-b547-2cbe178094d1-logs\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.995169 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qzbs\" (UniqueName: \"kubernetes.io/projected/17260561-5216-45fa-b547-2cbe178094d1-kube-api-access-2qzbs\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.995240 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.995351 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.995392 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.995467 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/17260561-5216-45fa-b547-2cbe178094d1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.995515 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17260561-5216-45fa-b547-2cbe178094d1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.995739 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:08 crc kubenswrapper[4912]: I1203 01:49:08.995777 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.098754 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qzbs\" (UniqueName: \"kubernetes.io/projected/17260561-5216-45fa-b547-2cbe178094d1-kube-api-access-2qzbs\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.098817 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.098865 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.098889 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.098913 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/17260561-5216-45fa-b547-2cbe178094d1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.098941 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17260561-5216-45fa-b547-2cbe178094d1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.099019 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.099038 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.099079 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17260561-5216-45fa-b547-2cbe178094d1-logs\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.101755 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/17260561-5216-45fa-b547-2cbe178094d1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.099659 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17260561-5216-45fa-b547-2cbe178094d1-logs\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.102795 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.390841 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.390864 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.391849 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.391867 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17260561-5216-45fa-b547-2cbe178094d1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.392803 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/17260561-5216-45fa-b547-2cbe178094d1-ceph\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.393141 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qzbs\" (UniqueName: \"kubernetes.io/projected/17260561-5216-45fa-b547-2cbe178094d1-kube-api-access-2qzbs\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.457627 4912 generic.go:334] "Generic (PLEG): container finished" podID="700218d5-259f-478b-94b3-abc87dc8c76f" containerID="76aefcb003ed1a03e9b7d4b58a3e9bdf63c9cec2ddcb6b5f9f90d8be255a0542" exitCode=0 Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.457658 4912 generic.go:334] "Generic (PLEG): container finished" podID="700218d5-259f-478b-94b3-abc87dc8c76f" containerID="4652d5b0bc1bd5458d9f5c40ead842bf1686b33f653f370fc6d8d50c4b4f8b65" exitCode=143 Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.457677 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"700218d5-259f-478b-94b3-abc87dc8c76f","Type":"ContainerDied","Data":"76aefcb003ed1a03e9b7d4b58a3e9bdf63c9cec2ddcb6b5f9f90d8be255a0542"} Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.457753 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"700218d5-259f-478b-94b3-abc87dc8c76f","Type":"ContainerDied","Data":"4652d5b0bc1bd5458d9f5c40ead842bf1686b33f653f370fc6d8d50c4b4f8b65"} Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.491465 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"17260561-5216-45fa-b547-2cbe178094d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.659360 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.690120 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.744853 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-logs\") pod \"700218d5-259f-478b-94b3-abc87dc8c76f\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.745168 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms55t\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-kube-api-access-ms55t\") pod \"700218d5-259f-478b-94b3-abc87dc8c76f\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.745246 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-config-data\") pod \"700218d5-259f-478b-94b3-abc87dc8c76f\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.745266 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-httpd-run\") pod \"700218d5-259f-478b-94b3-abc87dc8c76f\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.745313 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-scripts\") pod \"700218d5-259f-478b-94b3-abc87dc8c76f\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.745344 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-ceph\") pod \"700218d5-259f-478b-94b3-abc87dc8c76f\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.745386 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"700218d5-259f-478b-94b3-abc87dc8c76f\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.745452 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-combined-ca-bundle\") pod \"700218d5-259f-478b-94b3-abc87dc8c76f\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.745512 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-public-tls-certs\") pod \"700218d5-259f-478b-94b3-abc87dc8c76f\" (UID: \"700218d5-259f-478b-94b3-abc87dc8c76f\") " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.746353 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-logs" (OuterVolumeSpecName: "logs") pod "700218d5-259f-478b-94b3-abc87dc8c76f" (UID: "700218d5-259f-478b-94b3-abc87dc8c76f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.759446 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-kube-api-access-ms55t" (OuterVolumeSpecName: "kube-api-access-ms55t") pod "700218d5-259f-478b-94b3-abc87dc8c76f" (UID: "700218d5-259f-478b-94b3-abc87dc8c76f"). InnerVolumeSpecName "kube-api-access-ms55t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.761318 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "700218d5-259f-478b-94b3-abc87dc8c76f" (UID: "700218d5-259f-478b-94b3-abc87dc8c76f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.769277 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-ceph" (OuterVolumeSpecName: "ceph") pod "700218d5-259f-478b-94b3-abc87dc8c76f" (UID: "700218d5-259f-478b-94b3-abc87dc8c76f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.777097 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-scripts" (OuterVolumeSpecName: "scripts") pod "700218d5-259f-478b-94b3-abc87dc8c76f" (UID: "700218d5-259f-478b-94b3-abc87dc8c76f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.784630 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.785502 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.791580 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "700218d5-259f-478b-94b3-abc87dc8c76f" (UID: "700218d5-259f-478b-94b3-abc87dc8c76f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.803537 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "700218d5-259f-478b-94b3-abc87dc8c76f" (UID: "700218d5-259f-478b-94b3-abc87dc8c76f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.840523 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "700218d5-259f-478b-94b3-abc87dc8c76f" (UID: "700218d5-259f-478b-94b3-abc87dc8c76f"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.848787 4912 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.848815 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.848824 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.848842 4912 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.848851 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.848861 4912 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.848868 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700218d5-259f-478b-94b3-abc87dc8c76f-logs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.848879 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms55t\" (UniqueName: \"kubernetes.io/projected/700218d5-259f-478b-94b3-abc87dc8c76f-kube-api-access-ms55t\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.905283 4912 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.959792 4912 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:09 crc kubenswrapper[4912]: I1203 01:49:09.992288 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-config-data" (OuterVolumeSpecName: "config-data") pod "700218d5-259f-478b-94b3-abc87dc8c76f" (UID: "700218d5-259f-478b-94b3-abc87dc8c76f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.066304 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700218d5-259f-478b-94b3-abc87dc8c76f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.116386 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-h7pdq"] Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.459009 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.481096 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-h7pdq" event={"ID":"fe7da98f-555b-47d1-ae2a-849d69e7540a","Type":"ContainerStarted","Data":"6155b72a9ca6de375f77ae5059c2f47c1c32cdf7a562903d0f97691fc1a19d5e"} Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.484725 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.484731 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"700218d5-259f-478b-94b3-abc87dc8c76f","Type":"ContainerDied","Data":"7ad369b588d7ebf1f91c99647daf02911bfd09c96fb089ebae91399582fc9aab"} Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.484901 4912 scope.go:117] "RemoveContainer" containerID="76aefcb003ed1a03e9b7d4b58a3e9bdf63c9cec2ddcb6b5f9f90d8be255a0542" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.488609 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17260561-5216-45fa-b547-2cbe178094d1","Type":"ContainerStarted","Data":"a147873246ef2b56d73387a9f167135c8dd9aa5f01c598f779ff9aa449e1a1cb"} Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.528173 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.533585 4912 scope.go:117] "RemoveContainer" containerID="4652d5b0bc1bd5458d9f5c40ead842bf1686b33f653f370fc6d8d50c4b4f8b65" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.547621 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.563767 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:10 crc kubenswrapper[4912]: E1203 01:49:10.564843 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700218d5-259f-478b-94b3-abc87dc8c76f" containerName="glance-httpd" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.564860 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="700218d5-259f-478b-94b3-abc87dc8c76f" containerName="glance-httpd" Dec 03 01:49:10 crc kubenswrapper[4912]: E1203 01:49:10.564913 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="700218d5-259f-478b-94b3-abc87dc8c76f" containerName="glance-log" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.564921 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="700218d5-259f-478b-94b3-abc87dc8c76f" containerName="glance-log" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.565261 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="700218d5-259f-478b-94b3-abc87dc8c76f" containerName="glance-log" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.565304 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="700218d5-259f-478b-94b3-abc87dc8c76f" containerName="glance-httpd" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.571030 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.574533 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.574832 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.633565 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0846adb6-e9b8-4d98-ac3d-57fbf9d23a00" path="/var/lib/kubelet/pods/0846adb6-e9b8-4d98-ac3d-57fbf9d23a00/volumes" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.634368 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="700218d5-259f-478b-94b3-abc87dc8c76f" path="/var/lib/kubelet/pods/700218d5-259f-478b-94b3-abc87dc8c76f/volumes" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.635012 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.688612 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.688663 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf551edf-1401-42c2-8d75-6dec09a149c6-logs\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.688688 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.688799 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.688815 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf551edf-1401-42c2-8d75-6dec09a149c6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.688835 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bf551edf-1401-42c2-8d75-6dec09a149c6-ceph\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.688944 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjqlh\" (UniqueName: \"kubernetes.io/projected/bf551edf-1401-42c2-8d75-6dec09a149c6-kube-api-access-bjqlh\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.689008 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-config-data\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.689037 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-scripts\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.791314 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.791668 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf551edf-1401-42c2-8d75-6dec09a149c6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.791698 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bf551edf-1401-42c2-8d75-6dec09a149c6-ceph\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.791837 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjqlh\" (UniqueName: \"kubernetes.io/projected/bf551edf-1401-42c2-8d75-6dec09a149c6-kube-api-access-bjqlh\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.791902 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-config-data\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.791932 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-scripts\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.791979 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.792017 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf551edf-1401-42c2-8d75-6dec09a149c6-logs\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.792038 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.793500 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bf551edf-1401-42c2-8d75-6dec09a149c6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.793499 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.793539 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf551edf-1401-42c2-8d75-6dec09a149c6-logs\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:10 crc kubenswrapper[4912]: I1203 01:49:10.869585 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-s5kr2" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerName="registry-server" probeResult="failure" output=< Dec 03 01:49:10 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 01:49:10 crc kubenswrapper[4912]: > Dec 03 01:49:11 crc kubenswrapper[4912]: I1203 01:49:11.291332 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-config-data\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:11 crc kubenswrapper[4912]: I1203 01:49:11.291413 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:11 crc kubenswrapper[4912]: I1203 01:49:11.291845 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjqlh\" (UniqueName: \"kubernetes.io/projected/bf551edf-1401-42c2-8d75-6dec09a149c6-kube-api-access-bjqlh\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:11 crc kubenswrapper[4912]: I1203 01:49:11.292027 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:11 crc kubenswrapper[4912]: I1203 01:49:11.292532 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf551edf-1401-42c2-8d75-6dec09a149c6-scripts\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:11 crc kubenswrapper[4912]: I1203 01:49:11.292941 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bf551edf-1401-42c2-8d75-6dec09a149c6-ceph\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:11 crc kubenswrapper[4912]: I1203 01:49:11.343790 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"bf551edf-1401-42c2-8d75-6dec09a149c6\") " pod="openstack/glance-default-external-api-0" Dec 03 01:49:11 crc kubenswrapper[4912]: I1203 01:49:11.534172 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 01:49:12 crc kubenswrapper[4912]: I1203 01:49:12.154525 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 01:49:12 crc kubenswrapper[4912]: W1203 01:49:12.161866 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf551edf_1401_42c2_8d75_6dec09a149c6.slice/crio-75aec66b907fd8de51100361627ae114da18be29fcbae916552cd2fdf742a5ec WatchSource:0}: Error finding container 75aec66b907fd8de51100361627ae114da18be29fcbae916552cd2fdf742a5ec: Status 404 returned error can't find the container with id 75aec66b907fd8de51100361627ae114da18be29fcbae916552cd2fdf742a5ec Dec 03 01:49:12 crc kubenswrapper[4912]: I1203 01:49:12.421391 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 03 01:49:12 crc kubenswrapper[4912]: I1203 01:49:12.522294 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17260561-5216-45fa-b547-2cbe178094d1","Type":"ContainerStarted","Data":"731c701b28d8e971a106d6d6a6c6dc22aeeb0e88a85d6bfba0ef0ca9538b9ad1"} Dec 03 01:49:12 crc kubenswrapper[4912]: I1203 01:49:12.524472 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bf551edf-1401-42c2-8d75-6dec09a149c6","Type":"ContainerStarted","Data":"75aec66b907fd8de51100361627ae114da18be29fcbae916552cd2fdf742a5ec"} Dec 03 01:49:12 crc kubenswrapper[4912]: I1203 01:49:12.546132 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 03 01:49:13 crc kubenswrapper[4912]: I1203 01:49:13.544137 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"17260561-5216-45fa-b547-2cbe178094d1","Type":"ContainerStarted","Data":"c2a130729ba2c3f9fcef57f3647f3d98e7318ef1d82c3dd19428606e606bb76e"} Dec 03 01:49:13 crc kubenswrapper[4912]: I1203 01:49:13.573627 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.573610115 podStartE2EDuration="5.573610115s" podCreationTimestamp="2025-12-03 01:49:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:49:13.570235304 +0000 UTC m=+5139.212255864" watchObservedRunningTime="2025-12-03 01:49:13.573610115 +0000 UTC m=+5139.215630675" Dec 03 01:49:16 crc kubenswrapper[4912]: I1203 01:49:16.581552 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:49:16 crc kubenswrapper[4912]: E1203 01:49:16.582073 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:49:19 crc kubenswrapper[4912]: I1203 01:49:19.634478 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bf551edf-1401-42c2-8d75-6dec09a149c6","Type":"ContainerStarted","Data":"ab427d3678d45310764067d74f58d88106a686d7287814c973456a62abe06a43"} Dec 03 01:49:19 crc kubenswrapper[4912]: I1203 01:49:19.659688 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:19 crc kubenswrapper[4912]: I1203 01:49:19.660886 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:19 crc kubenswrapper[4912]: I1203 01:49:19.699560 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:19 crc kubenswrapper[4912]: I1203 01:49:19.711111 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:19 crc kubenswrapper[4912]: I1203 01:49:19.833128 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:49:19 crc kubenswrapper[4912]: I1203 01:49:19.882575 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:49:20 crc kubenswrapper[4912]: I1203 01:49:20.066421 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s5kr2"] Dec 03 01:49:20 crc kubenswrapper[4912]: I1203 01:49:20.644739 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:20 crc kubenswrapper[4912]: I1203 01:49:20.644782 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:21 crc kubenswrapper[4912]: I1203 01:49:21.662021 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s5kr2" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerName="registry-server" containerID="cri-o://4c1ebe643ab6278e030ac9d71815b994375f9bb5d498c8c351f80604b0deec72" gracePeriod=2 Dec 03 01:49:22 crc kubenswrapper[4912]: I1203 01:49:22.707391 4912 generic.go:334] "Generic (PLEG): container finished" podID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerID="4c1ebe643ab6278e030ac9d71815b994375f9bb5d498c8c351f80604b0deec72" exitCode=0 Dec 03 01:49:22 crc kubenswrapper[4912]: I1203 01:49:22.707908 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5kr2" event={"ID":"96532d70-09e3-4830-a4e7-1349fb2e602a","Type":"ContainerDied","Data":"4c1ebe643ab6278e030ac9d71815b994375f9bb5d498c8c351f80604b0deec72"} Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.082574 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.227871 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-catalog-content\") pod \"96532d70-09e3-4830-a4e7-1349fb2e602a\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.228608 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv8mh\" (UniqueName: \"kubernetes.io/projected/96532d70-09e3-4830-a4e7-1349fb2e602a-kube-api-access-nv8mh\") pod \"96532d70-09e3-4830-a4e7-1349fb2e602a\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.228824 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-utilities\") pod \"96532d70-09e3-4830-a4e7-1349fb2e602a\" (UID: \"96532d70-09e3-4830-a4e7-1349fb2e602a\") " Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.229367 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-utilities" (OuterVolumeSpecName: "utilities") pod "96532d70-09e3-4830-a4e7-1349fb2e602a" (UID: "96532d70-09e3-4830-a4e7-1349fb2e602a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.231071 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.235020 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96532d70-09e3-4830-a4e7-1349fb2e602a-kube-api-access-nv8mh" (OuterVolumeSpecName: "kube-api-access-nv8mh") pod "96532d70-09e3-4830-a4e7-1349fb2e602a" (UID: "96532d70-09e3-4830-a4e7-1349fb2e602a"). InnerVolumeSpecName "kube-api-access-nv8mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.289178 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96532d70-09e3-4830-a4e7-1349fb2e602a" (UID: "96532d70-09e3-4830-a4e7-1349fb2e602a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.332215 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96532d70-09e3-4830-a4e7-1349fb2e602a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.332255 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv8mh\" (UniqueName: \"kubernetes.io/projected/96532d70-09e3-4830-a4e7-1349fb2e602a-kube-api-access-nv8mh\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.723183 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77dc7f54bf-f2rhc" event={"ID":"a6880b8c-c092-43bd-9cd1-8c837be9bfe2","Type":"ContainerStarted","Data":"f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87"} Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.727214 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s5kr2" event={"ID":"96532d70-09e3-4830-a4e7-1349fb2e602a","Type":"ContainerDied","Data":"c376e4ec8e13f8919ddee2d09beec5efaa12682017dd3a4646d21462b1b7874e"} Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.727256 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s5kr2" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.727275 4912 scope.go:117] "RemoveContainer" containerID="4c1ebe643ab6278e030ac9d71815b994375f9bb5d498c8c351f80604b0deec72" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.729169 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4dbbf8d-5vgmn" event={"ID":"3d92b6a2-0079-4d39-9c11-9a31c22429f4","Type":"ContainerStarted","Data":"5ca3c6907f42a840274852105bab49409bf2ce1662393731ebac52ad3bb19db2"} Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.733868 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d9dc8d44b-7tljg" event={"ID":"08762529-e1c5-4358-bc43-c0d81a4c6df6","Type":"ContainerStarted","Data":"c77922fad380b1690cfe80e8b1b9cb7b6921b63e8a0da1cdc0b8624aac711282"} Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.736209 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.736291 4912 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.739268 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 01:49:23 crc kubenswrapper[4912]: I1203 01:49:23.739837 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69f9d58fc9-6l6qj" event={"ID":"7f9ca12c-238d-47f3-a39d-6c2df17d6baa","Type":"ContainerStarted","Data":"cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512"} Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.246026 4912 scope.go:117] "RemoveContainer" containerID="8c0022724bb5acd03361b37a668e5e0049cb7db1f2a821d9ae63dbdc13822123" Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.292206 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s5kr2"] Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.306795 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s5kr2"] Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.326551 4912 scope.go:117] "RemoveContainer" containerID="23b03489859f570820bd666c45f9d53175721d55978ca02df11f21c9c1ccc16f" Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.590825 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" path="/var/lib/kubelet/pods/96532d70-09e3-4830-a4e7-1349fb2e602a/volumes" Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.752077 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d4dbbf8d-5vgmn" event={"ID":"3d92b6a2-0079-4d39-9c11-9a31c22429f4","Type":"ContainerStarted","Data":"6536f14cc79f4069da9ee8ff8ea6deaf5e43dd852b26c4da51b7fe726ac3509a"} Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.755710 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d9dc8d44b-7tljg" event={"ID":"08762529-e1c5-4358-bc43-c0d81a4c6df6","Type":"ContainerStarted","Data":"d6aa5aaf08c7d099e3be624e725ccb79ec7adb51678136411bc203ceed6842ea"} Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.757550 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69f9d58fc9-6l6qj" event={"ID":"7f9ca12c-238d-47f3-a39d-6c2df17d6baa","Type":"ContainerStarted","Data":"dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785"} Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.757604 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69f9d58fc9-6l6qj" podUID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerName="horizon-log" containerID="cri-o://cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512" gracePeriod=30 Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.757645 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-69f9d58fc9-6l6qj" podUID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerName="horizon" containerID="cri-o://dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785" gracePeriod=30 Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.766959 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bf551edf-1401-42c2-8d75-6dec09a149c6","Type":"ContainerStarted","Data":"e92621ccb7b2a91e3ce5309a9be5f4cd1e454a00af916c78f06222e2777d59cb"} Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.771786 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77dc7f54bf-f2rhc" podUID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerName="horizon-log" containerID="cri-o://f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87" gracePeriod=30 Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.771808 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77dc7f54bf-f2rhc" event={"ID":"a6880b8c-c092-43bd-9cd1-8c837be9bfe2","Type":"ContainerStarted","Data":"12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9"} Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.771969 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77dc7f54bf-f2rhc" podUID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerName="horizon" containerID="cri-o://12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9" gracePeriod=30 Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.790823 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-h7pdq" event={"ID":"fe7da98f-555b-47d1-ae2a-849d69e7540a","Type":"ContainerStarted","Data":"2e164f6062c79afe348e86fcd5e0dabe2d990d246afd65719452b71c35028c7b"} Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.796340 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d4dbbf8d-5vgmn" podStartSLOduration=3.566738847 podStartE2EDuration="18.796321979s" podCreationTimestamp="2025-12-03 01:49:06 +0000 UTC" firstStartedPulling="2025-12-03 01:49:07.496705768 +0000 UTC m=+5133.138726328" lastFinishedPulling="2025-12-03 01:49:22.7262889 +0000 UTC m=+5148.368309460" observedRunningTime="2025-12-03 01:49:24.779021763 +0000 UTC m=+5150.421042333" watchObservedRunningTime="2025-12-03 01:49:24.796321979 +0000 UTC m=+5150.438342539" Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.817189 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-77dc7f54bf-f2rhc" podStartSLOduration=4.333748072 podStartE2EDuration="22.817170461s" podCreationTimestamp="2025-12-03 01:49:02 +0000 UTC" firstStartedPulling="2025-12-03 01:49:04.247933378 +0000 UTC m=+5129.889953938" lastFinishedPulling="2025-12-03 01:49:22.731355767 +0000 UTC m=+5148.373376327" observedRunningTime="2025-12-03 01:49:24.808857587 +0000 UTC m=+5150.450878157" watchObservedRunningTime="2025-12-03 01:49:24.817170461 +0000 UTC m=+5150.459191021" Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.841674 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=14.84165676 podStartE2EDuration="14.84165676s" podCreationTimestamp="2025-12-03 01:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:49:24.831315862 +0000 UTC m=+5150.473336412" watchObservedRunningTime="2025-12-03 01:49:24.84165676 +0000 UTC m=+5150.483677320" Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.853917 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-69f9d58fc9-6l6qj" podStartSLOduration=4.171094038 podStartE2EDuration="22.85389662s" podCreationTimestamp="2025-12-03 01:49:02 +0000 UTC" firstStartedPulling="2025-12-03 01:49:04.04947342 +0000 UTC m=+5129.691493980" lastFinishedPulling="2025-12-03 01:49:22.732276002 +0000 UTC m=+5148.374296562" observedRunningTime="2025-12-03 01:49:24.84722723 +0000 UTC m=+5150.489247790" watchObservedRunningTime="2025-12-03 01:49:24.85389662 +0000 UTC m=+5150.495917180" Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.871539 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-d9dc8d44b-7tljg" podStartSLOduration=4.491472639 podStartE2EDuration="19.871518686s" podCreationTimestamp="2025-12-03 01:49:05 +0000 UTC" firstStartedPulling="2025-12-03 01:49:07.355281857 +0000 UTC m=+5132.997302417" lastFinishedPulling="2025-12-03 01:49:22.735327904 +0000 UTC m=+5148.377348464" observedRunningTime="2025-12-03 01:49:24.864569048 +0000 UTC m=+5150.506589618" watchObservedRunningTime="2025-12-03 01:49:24.871518686 +0000 UTC m=+5150.513539246" Dec 03 01:49:24 crc kubenswrapper[4912]: I1203 01:49:24.887635 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-h7pdq" podStartSLOduration=4.277597722 podStartE2EDuration="16.887617839s" podCreationTimestamp="2025-12-03 01:49:08 +0000 UTC" firstStartedPulling="2025-12-03 01:49:10.123293083 +0000 UTC m=+5135.765313643" lastFinishedPulling="2025-12-03 01:49:22.7333132 +0000 UTC m=+5148.375333760" observedRunningTime="2025-12-03 01:49:24.878837133 +0000 UTC m=+5150.520857713" watchObservedRunningTime="2025-12-03 01:49:24.887617839 +0000 UTC m=+5150.529638399" Dec 03 01:49:26 crc kubenswrapper[4912]: I1203 01:49:26.340022 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:26 crc kubenswrapper[4912]: I1203 01:49:26.340078 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:26 crc kubenswrapper[4912]: I1203 01:49:26.606959 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:26 crc kubenswrapper[4912]: I1203 01:49:26.607667 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:28 crc kubenswrapper[4912]: I1203 01:49:28.572645 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:49:28 crc kubenswrapper[4912]: E1203 01:49:28.573322 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:49:31 crc kubenswrapper[4912]: I1203 01:49:31.535484 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 01:49:31 crc kubenswrapper[4912]: I1203 01:49:31.536160 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 01:49:32 crc kubenswrapper[4912]: I1203 01:49:32.049572 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 01:49:32 crc kubenswrapper[4912]: I1203 01:49:32.050092 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 01:49:32 crc kubenswrapper[4912]: I1203 01:49:32.064927 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 01:49:32 crc kubenswrapper[4912]: I1203 01:49:32.877542 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 01:49:33 crc kubenswrapper[4912]: I1203 01:49:33.316665 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:33 crc kubenswrapper[4912]: I1203 01:49:33.646585 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:33 crc kubenswrapper[4912]: I1203 01:49:33.887006 4912 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 01:49:34 crc kubenswrapper[4912]: I1203 01:49:34.377290 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 01:49:34 crc kubenswrapper[4912]: I1203 01:49:34.379931 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 01:49:36 crc kubenswrapper[4912]: I1203 01:49:36.343337 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-d9dc8d44b-7tljg" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.69:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.69:8443: connect: connection refused" Dec 03 01:49:36 crc kubenswrapper[4912]: I1203 01:49:36.607813 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6d4dbbf8d-5vgmn" podUID="3d92b6a2-0079-4d39-9c11-9a31c22429f4" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.70:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.70:8443: connect: connection refused" Dec 03 01:49:36 crc kubenswrapper[4912]: I1203 01:49:36.914897 4912 generic.go:334] "Generic (PLEG): container finished" podID="fe7da98f-555b-47d1-ae2a-849d69e7540a" containerID="2e164f6062c79afe348e86fcd5e0dabe2d990d246afd65719452b71c35028c7b" exitCode=0 Dec 03 01:49:36 crc kubenswrapper[4912]: I1203 01:49:36.914949 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-h7pdq" event={"ID":"fe7da98f-555b-47d1-ae2a-849d69e7540a","Type":"ContainerDied","Data":"2e164f6062c79afe348e86fcd5e0dabe2d990d246afd65719452b71c35028c7b"} Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.654595 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.735522 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-config-data\") pod \"fe7da98f-555b-47d1-ae2a-849d69e7540a\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.735866 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-job-config-data\") pod \"fe7da98f-555b-47d1-ae2a-849d69e7540a\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.735907 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wglrw\" (UniqueName: \"kubernetes.io/projected/fe7da98f-555b-47d1-ae2a-849d69e7540a-kube-api-access-wglrw\") pod \"fe7da98f-555b-47d1-ae2a-849d69e7540a\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.736028 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-combined-ca-bundle\") pod \"fe7da98f-555b-47d1-ae2a-849d69e7540a\" (UID: \"fe7da98f-555b-47d1-ae2a-849d69e7540a\") " Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.746252 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "fe7da98f-555b-47d1-ae2a-849d69e7540a" (UID: "fe7da98f-555b-47d1-ae2a-849d69e7540a"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.758552 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe7da98f-555b-47d1-ae2a-849d69e7540a-kube-api-access-wglrw" (OuterVolumeSpecName: "kube-api-access-wglrw") pod "fe7da98f-555b-47d1-ae2a-849d69e7540a" (UID: "fe7da98f-555b-47d1-ae2a-849d69e7540a"). InnerVolumeSpecName "kube-api-access-wglrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.763422 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-config-data" (OuterVolumeSpecName: "config-data") pod "fe7da98f-555b-47d1-ae2a-849d69e7540a" (UID: "fe7da98f-555b-47d1-ae2a-849d69e7540a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.791642 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe7da98f-555b-47d1-ae2a-849d69e7540a" (UID: "fe7da98f-555b-47d1-ae2a-849d69e7540a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.838483 4912 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.838517 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wglrw\" (UniqueName: \"kubernetes.io/projected/fe7da98f-555b-47d1-ae2a-849d69e7540a-kube-api-access-wglrw\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.838528 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.838536 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7da98f-555b-47d1-ae2a-849d69e7540a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.947987 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-h7pdq" event={"ID":"fe7da98f-555b-47d1-ae2a-849d69e7540a","Type":"ContainerDied","Data":"6155b72a9ca6de375f77ae5059c2f47c1c32cdf7a562903d0f97691fc1a19d5e"} Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.948026 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6155b72a9ca6de375f77ae5059c2f47c1c32cdf7a562903d0f97691fc1a19d5e" Dec 03 01:49:38 crc kubenswrapper[4912]: I1203 01:49:38.948051 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-h7pdq" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.313505 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 01:49:39 crc kubenswrapper[4912]: E1203 01:49:39.314034 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerName="registry-server" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.314051 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerName="registry-server" Dec 03 01:49:39 crc kubenswrapper[4912]: E1203 01:49:39.314070 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerName="extract-content" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.314078 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerName="extract-content" Dec 03 01:49:39 crc kubenswrapper[4912]: E1203 01:49:39.314091 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7da98f-555b-47d1-ae2a-849d69e7540a" containerName="manila-db-sync" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.314098 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7da98f-555b-47d1-ae2a-849d69e7540a" containerName="manila-db-sync" Dec 03 01:49:39 crc kubenswrapper[4912]: E1203 01:49:39.314112 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerName="extract-utilities" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.314118 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerName="extract-utilities" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.314363 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7da98f-555b-47d1-ae2a-849d69e7540a" containerName="manila-db-sync" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.314387 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="96532d70-09e3-4830-a4e7-1349fb2e602a" containerName="registry-server" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.315634 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.318944 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-4ttg5" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.319207 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.320055 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.323245 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.350690 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.350743 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6xsz\" (UniqueName: \"kubernetes.io/projected/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-kube-api-access-v6xsz\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.350804 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.350839 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.350911 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.350957 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-scripts\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.361839 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.433551 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.435516 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.439348 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.447499 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c8d8d886c-p2gz6"] Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.449390 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.458156 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.458234 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-scripts\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.458286 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.458311 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6xsz\" (UniqueName: \"kubernetes.io/projected/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-kube-api-access-v6xsz\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.458364 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.458395 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.465490 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.470190 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-scripts\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.486475 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.487355 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.507448 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.510911 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6xsz\" (UniqueName: \"kubernetes.io/projected/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-kube-api-access-v6xsz\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.519661 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c8d8d886c-p2gz6"] Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.528204 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.567900 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.567993 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-468ql\" (UniqueName: \"kubernetes.io/projected/51b6bb89-372c-4193-ab71-6317807c1820-kube-api-access-468ql\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568054 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-dns-svc\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568091 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-dns-swift-storage-0\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568117 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmv4f\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-kube-api-access-nmv4f\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568219 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568403 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568459 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-scripts\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568484 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568532 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-ceph\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568607 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-ovsdbserver-sb\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568842 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568919 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-ovsdbserver-nb\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568979 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-config\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.568996 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-openstack-edpm-ipam\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.629233 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.631496 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.646079 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.648799 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.668052 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.671175 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-ovsdbserver-sb\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.675556 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data-custom\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.676998 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.677093 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-ovsdbserver-nb\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.677182 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-openstack-edpm-ipam\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.677255 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-config\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.677368 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-scripts\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.673731 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-ovsdbserver-sb\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.677737 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc963855-e556-46df-a8b0-d3a9a94bdd9e-etc-machine-id\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.680838 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.681740 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.681922 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-468ql\" (UniqueName: \"kubernetes.io/projected/51b6bb89-372c-4193-ab71-6317807c1820-kube-api-access-468ql\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.679934 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-openstack-edpm-ipam\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.679403 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-config\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682036 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682476 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-dns-svc\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682544 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-dns-swift-storage-0\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682580 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmv4f\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-kube-api-access-nmv4f\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682612 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggvqf\" (UniqueName: \"kubernetes.io/projected/cc963855-e556-46df-a8b0-d3a9a94bdd9e-kube-api-access-ggvqf\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682638 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682762 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc963855-e556-46df-a8b0-d3a9a94bdd9e-logs\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682808 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682837 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-scripts\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682867 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.682913 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-ceph\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.687416 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-ceph\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.687582 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.688949 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.689508 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-dns-svc\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.679576 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.689620 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-dns-swift-storage-0\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.680195 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51b6bb89-372c-4193-ab71-6317807c1820-ovsdbserver-nb\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.696825 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.697209 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-scripts\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.702126 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.714938 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-468ql\" (UniqueName: \"kubernetes.io/projected/51b6bb89-372c-4193-ab71-6317807c1820-kube-api-access-468ql\") pod \"dnsmasq-dns-c8d8d886c-p2gz6\" (UID: \"51b6bb89-372c-4193-ab71-6317807c1820\") " pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.718738 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmv4f\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-kube-api-access-nmv4f\") pod \"manila-share-share1-0\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.785459 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.785720 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.785795 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggvqf\" (UniqueName: \"kubernetes.io/projected/cc963855-e556-46df-a8b0-d3a9a94bdd9e-kube-api-access-ggvqf\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.785849 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc963855-e556-46df-a8b0-d3a9a94bdd9e-logs\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.786473 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc963855-e556-46df-a8b0-d3a9a94bdd9e-logs\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.787876 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data-custom\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.788105 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc963855-e556-46df-a8b0-d3a9a94bdd9e-etc-machine-id\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.788120 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-scripts\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.793310 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.793406 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc963855-e556-46df-a8b0-d3a9a94bdd9e-etc-machine-id\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.793686 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-scripts\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.794381 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data-custom\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.804165 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.815994 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggvqf\" (UniqueName: \"kubernetes.io/projected/cc963855-e556-46df-a8b0-d3a9a94bdd9e-kube-api-access-ggvqf\") pod \"manila-api-0\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " pod="openstack/manila-api-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.898454 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.925089 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:39 crc kubenswrapper[4912]: I1203 01:49:39.985597 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 01:49:40 crc kubenswrapper[4912]: I1203 01:49:40.450796 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 01:49:40 crc kubenswrapper[4912]: I1203 01:49:40.464638 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:49:40 crc kubenswrapper[4912]: I1203 01:49:40.678848 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c8d8d886c-p2gz6"] Dec 03 01:49:41 crc kubenswrapper[4912]: I1203 01:49:41.046664 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 01:49:41 crc kubenswrapper[4912]: I1203 01:49:41.048021 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" event={"ID":"51b6bb89-372c-4193-ab71-6317807c1820","Type":"ContainerStarted","Data":"b32f52bc3e58bdd849a72dc3c7012e74f377f239f3f43df2637e9e281b198f84"} Dec 03 01:49:41 crc kubenswrapper[4912]: I1203 01:49:41.053685 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1","Type":"ContainerStarted","Data":"5142ee9482a19b7e3791df7ee0548a2293eed18123eba8baec80ac00a735bdd0"} Dec 03 01:49:41 crc kubenswrapper[4912]: I1203 01:49:41.198595 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 01:49:42 crc kubenswrapper[4912]: I1203 01:49:42.067493 4912 generic.go:334] "Generic (PLEG): container finished" podID="51b6bb89-372c-4193-ab71-6317807c1820" containerID="456a0213b9d679dc400d5cec259753c7e4579237d0b4da55c0174f0c15162d9f" exitCode=0 Dec 03 01:49:42 crc kubenswrapper[4912]: I1203 01:49:42.068740 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" event={"ID":"51b6bb89-372c-4193-ab71-6317807c1820","Type":"ContainerDied","Data":"456a0213b9d679dc400d5cec259753c7e4579237d0b4da55c0174f0c15162d9f"} Dec 03 01:49:42 crc kubenswrapper[4912]: I1203 01:49:42.075711 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cc963855-e556-46df-a8b0-d3a9a94bdd9e","Type":"ContainerStarted","Data":"aad25c7c51340b95c7f557ac9406731c96c2aed92e3ff1db0c7c3a166ff19c4a"} Dec 03 01:49:42 crc kubenswrapper[4912]: I1203 01:49:42.080178 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"163bd5c4-1469-4cdd-b39f-00bf454a0adb","Type":"ContainerStarted","Data":"2eee6918fa22c516383e9e6eb3361b197df8e7db124b3e659598cf7f331398fd"} Dec 03 01:49:42 crc kubenswrapper[4912]: I1203 01:49:42.579182 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:49:42 crc kubenswrapper[4912]: E1203 01:49:42.579916 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:49:43 crc kubenswrapper[4912]: I1203 01:49:43.129541 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cc963855-e556-46df-a8b0-d3a9a94bdd9e","Type":"ContainerStarted","Data":"5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff"} Dec 03 01:49:43 crc kubenswrapper[4912]: I1203 01:49:43.132033 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1","Type":"ContainerStarted","Data":"6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3"} Dec 03 01:49:43 crc kubenswrapper[4912]: I1203 01:49:43.348980 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 03 01:49:44 crc kubenswrapper[4912]: I1203 01:49:44.151200 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" event={"ID":"51b6bb89-372c-4193-ab71-6317807c1820","Type":"ContainerStarted","Data":"e47df6f7e6ed264b2165581db68c300c8b3ca998d60b918c13b284ec176a02f3"} Dec 03 01:49:44 crc kubenswrapper[4912]: I1203 01:49:44.151604 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:44 crc kubenswrapper[4912]: I1203 01:49:44.166031 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cc963855-e556-46df-a8b0-d3a9a94bdd9e","Type":"ContainerStarted","Data":"15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf"} Dec 03 01:49:44 crc kubenswrapper[4912]: I1203 01:49:44.166172 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerName="manila-api-log" containerID="cri-o://5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff" gracePeriod=30 Dec 03 01:49:44 crc kubenswrapper[4912]: I1203 01:49:44.166218 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 03 01:49:44 crc kubenswrapper[4912]: I1203 01:49:44.166254 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerName="manila-api" containerID="cri-o://15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf" gracePeriod=30 Dec 03 01:49:44 crc kubenswrapper[4912]: I1203 01:49:44.205727 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" podStartSLOduration=5.205709558 podStartE2EDuration="5.205709558s" podCreationTimestamp="2025-12-03 01:49:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:49:44.194695831 +0000 UTC m=+5169.836716401" watchObservedRunningTime="2025-12-03 01:49:44.205709558 +0000 UTC m=+5169.847730118" Dec 03 01:49:44 crc kubenswrapper[4912]: I1203 01:49:44.231716 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=5.231695849 podStartE2EDuration="5.231695849s" podCreationTimestamp="2025-12-03 01:49:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:49:44.219951163 +0000 UTC m=+5169.861971723" watchObservedRunningTime="2025-12-03 01:49:44.231695849 +0000 UTC m=+5169.873716409" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.125961 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.210187 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc963855-e556-46df-a8b0-d3a9a94bdd9e-logs\") pod \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.210275 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-scripts\") pod \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.210298 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc963855-e556-46df-a8b0-d3a9a94bdd9e-etc-machine-id\") pod \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.210345 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-combined-ca-bundle\") pod \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.210417 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data\") pod \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.210531 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data-custom\") pod \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.210612 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggvqf\" (UniqueName: \"kubernetes.io/projected/cc963855-e556-46df-a8b0-d3a9a94bdd9e-kube-api-access-ggvqf\") pod \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\" (UID: \"cc963855-e556-46df-a8b0-d3a9a94bdd9e\") " Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.211023 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc963855-e556-46df-a8b0-d3a9a94bdd9e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cc963855-e556-46df-a8b0-d3a9a94bdd9e" (UID: "cc963855-e556-46df-a8b0-d3a9a94bdd9e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.211378 4912 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cc963855-e556-46df-a8b0-d3a9a94bdd9e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.220885 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc963855-e556-46df-a8b0-d3a9a94bdd9e-logs" (OuterVolumeSpecName: "logs") pod "cc963855-e556-46df-a8b0-d3a9a94bdd9e" (UID: "cc963855-e556-46df-a8b0-d3a9a94bdd9e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.227623 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cc963855-e556-46df-a8b0-d3a9a94bdd9e" (UID: "cc963855-e556-46df-a8b0-d3a9a94bdd9e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.230622 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc963855-e556-46df-a8b0-d3a9a94bdd9e-kube-api-access-ggvqf" (OuterVolumeSpecName: "kube-api-access-ggvqf") pod "cc963855-e556-46df-a8b0-d3a9a94bdd9e" (UID: "cc963855-e556-46df-a8b0-d3a9a94bdd9e"). InnerVolumeSpecName "kube-api-access-ggvqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.230691 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-scripts" (OuterVolumeSpecName: "scripts") pod "cc963855-e556-46df-a8b0-d3a9a94bdd9e" (UID: "cc963855-e556-46df-a8b0-d3a9a94bdd9e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.254493 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1","Type":"ContainerStarted","Data":"8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14"} Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.292098 4912 generic.go:334] "Generic (PLEG): container finished" podID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerID="15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf" exitCode=143 Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.292131 4912 generic.go:334] "Generic (PLEG): container finished" podID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerID="5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff" exitCode=143 Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.293210 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.293397 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cc963855-e556-46df-a8b0-d3a9a94bdd9e","Type":"ContainerDied","Data":"15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf"} Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.293451 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cc963855-e556-46df-a8b0-d3a9a94bdd9e","Type":"ContainerDied","Data":"5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff"} Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.293465 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"cc963855-e556-46df-a8b0-d3a9a94bdd9e","Type":"ContainerDied","Data":"aad25c7c51340b95c7f557ac9406731c96c2aed92e3ff1db0c7c3a166ff19c4a"} Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.293481 4912 scope.go:117] "RemoveContainer" containerID="15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.310514 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc963855-e556-46df-a8b0-d3a9a94bdd9e" (UID: "cc963855-e556-46df-a8b0-d3a9a94bdd9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.313138 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc963855-e556-46df-a8b0-d3a9a94bdd9e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.313167 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.313176 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.313185 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.313193 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggvqf\" (UniqueName: \"kubernetes.io/projected/cc963855-e556-46df-a8b0-d3a9a94bdd9e-kube-api-access-ggvqf\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.315086 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=5.489204987 podStartE2EDuration="6.315072111s" podCreationTimestamp="2025-12-03 01:49:39 +0000 UTC" firstStartedPulling="2025-12-03 01:49:40.464453377 +0000 UTC m=+5166.106473937" lastFinishedPulling="2025-12-03 01:49:41.290320501 +0000 UTC m=+5166.932341061" observedRunningTime="2025-12-03 01:49:45.297767914 +0000 UTC m=+5170.939788474" watchObservedRunningTime="2025-12-03 01:49:45.315072111 +0000 UTC m=+5170.957092671" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.364604 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data" (OuterVolumeSpecName: "config-data") pod "cc963855-e556-46df-a8b0-d3a9a94bdd9e" (UID: "cc963855-e556-46df-a8b0-d3a9a94bdd9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.414915 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc963855-e556-46df-a8b0-d3a9a94bdd9e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.456806 4912 scope.go:117] "RemoveContainer" containerID="5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.486618 4912 scope.go:117] "RemoveContainer" containerID="15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf" Dec 03 01:49:45 crc kubenswrapper[4912]: E1203 01:49:45.487025 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf\": container with ID starting with 15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf not found: ID does not exist" containerID="15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.487063 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf"} err="failed to get container status \"15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf\": rpc error: code = NotFound desc = could not find container \"15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf\": container with ID starting with 15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf not found: ID does not exist" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.487087 4912 scope.go:117] "RemoveContainer" containerID="5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff" Dec 03 01:49:45 crc kubenswrapper[4912]: E1203 01:49:45.487556 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff\": container with ID starting with 5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff not found: ID does not exist" containerID="5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.487592 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff"} err="failed to get container status \"5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff\": rpc error: code = NotFound desc = could not find container \"5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff\": container with ID starting with 5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff not found: ID does not exist" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.487614 4912 scope.go:117] "RemoveContainer" containerID="15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.487842 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf"} err="failed to get container status \"15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf\": rpc error: code = NotFound desc = could not find container \"15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf\": container with ID starting with 15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf not found: ID does not exist" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.487870 4912 scope.go:117] "RemoveContainer" containerID="5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.488163 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff"} err="failed to get container status \"5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff\": rpc error: code = NotFound desc = could not find container \"5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff\": container with ID starting with 5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff not found: ID does not exist" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.651485 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.664744 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.682353 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 03 01:49:45 crc kubenswrapper[4912]: E1203 01:49:45.682804 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerName="manila-api-log" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.682820 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerName="manila-api-log" Dec 03 01:49:45 crc kubenswrapper[4912]: E1203 01:49:45.682846 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerName="manila-api" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.682852 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerName="manila-api" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.683054 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerName="manila-api-log" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.683073 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" containerName="manila-api" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.684205 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.689894 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.689934 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.690059 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.697200 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.821839 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-etc-machine-id\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.821972 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-config-data-custom\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.822004 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-config-data\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.822095 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rckwb\" (UniqueName: \"kubernetes.io/projected/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-kube-api-access-rckwb\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.822126 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-logs\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.822168 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-internal-tls-certs\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.822194 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-scripts\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.822222 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-public-tls-certs\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.822268 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925009 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rckwb\" (UniqueName: \"kubernetes.io/projected/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-kube-api-access-rckwb\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925321 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-logs\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925371 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-internal-tls-certs\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925399 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-scripts\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925423 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-public-tls-certs\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925485 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925545 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-etc-machine-id\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925600 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-config-data-custom\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925623 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-config-data\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925812 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-logs\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.925881 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-etc-machine-id\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.930884 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.931551 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-internal-tls-certs\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.932467 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-config-data-custom\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.934193 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-scripts\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.936862 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-public-tls-certs\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.943096 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-config-data\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:45 crc kubenswrapper[4912]: I1203 01:49:45.968090 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rckwb\" (UniqueName: \"kubernetes.io/projected/eedcff7d-7bb2-4c30-8166-e7a2f22ac60e-kube-api-access-rckwb\") pod \"manila-api-0\" (UID: \"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e\") " pod="openstack/manila-api-0" Dec 03 01:49:46 crc kubenswrapper[4912]: I1203 01:49:46.029099 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 01:49:46 crc kubenswrapper[4912]: I1203 01:49:46.341911 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-d9dc8d44b-7tljg" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.69:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.69:8443: connect: connection refused" Dec 03 01:49:46 crc kubenswrapper[4912]: I1203 01:49:46.591500 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc963855-e556-46df-a8b0-d3a9a94bdd9e" path="/var/lib/kubelet/pods/cc963855-e556-46df-a8b0-d3a9a94bdd9e/volumes" Dec 03 01:49:46 crc kubenswrapper[4912]: I1203 01:49:46.606625 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6d4dbbf8d-5vgmn" podUID="3d92b6a2-0079-4d39-9c11-9a31c22429f4" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.70:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.70:8443: connect: connection refused" Dec 03 01:49:46 crc kubenswrapper[4912]: I1203 01:49:46.744725 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 01:49:47 crc kubenswrapper[4912]: I1203 01:49:47.358780 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e","Type":"ContainerStarted","Data":"43604f2f57aebbea2e51a8952d3121282c090f9b5572f3f5be8924f0cb85e7aa"} Dec 03 01:49:47 crc kubenswrapper[4912]: I1203 01:49:47.359284 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e","Type":"ContainerStarted","Data":"af499cf96e08416e34713ce396fff40c76e6991e777be845eccde0f9edc45c9f"} Dec 03 01:49:48 crc kubenswrapper[4912]: I1203 01:49:48.382914 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"eedcff7d-7bb2-4c30-8166-e7a2f22ac60e","Type":"ContainerStarted","Data":"e4c6c2901591a03f7ad7c46f1da8744c82a2df7b01a0b376739213b880d7df66"} Dec 03 01:49:48 crc kubenswrapper[4912]: I1203 01:49:48.384007 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 03 01:49:48 crc kubenswrapper[4912]: I1203 01:49:48.427972 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.42795472 podStartE2EDuration="3.42795472s" podCreationTimestamp="2025-12-03 01:49:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:49:48.421876137 +0000 UTC m=+5174.063896707" watchObservedRunningTime="2025-12-03 01:49:48.42795472 +0000 UTC m=+5174.069975280" Dec 03 01:49:49 crc kubenswrapper[4912]: I1203 01:49:49.669066 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 03 01:49:49 crc kubenswrapper[4912]: I1203 01:49:49.927917 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c8d8d886c-p2gz6" Dec 03 01:49:50 crc kubenswrapper[4912]: I1203 01:49:50.013175 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf"] Dec 03 01:49:50 crc kubenswrapper[4912]: I1203 01:49:50.013409 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" podUID="d819df61-e2e4-4960-8ff8-761a99f21850" containerName="dnsmasq-dns" containerID="cri-o://3cd1d24a28b305d242e061d5438fee63c1ba6ebc7f164d9027689fc893781107" gracePeriod=10 Dec 03 01:49:50 crc kubenswrapper[4912]: I1203 01:49:50.411875 4912 generic.go:334] "Generic (PLEG): container finished" podID="d819df61-e2e4-4960-8ff8-761a99f21850" containerID="3cd1d24a28b305d242e061d5438fee63c1ba6ebc7f164d9027689fc893781107" exitCode=0 Dec 03 01:49:50 crc kubenswrapper[4912]: I1203 01:49:50.411924 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" event={"ID":"d819df61-e2e4-4960-8ff8-761a99f21850","Type":"ContainerDied","Data":"3cd1d24a28b305d242e061d5438fee63c1ba6ebc7f164d9027689fc893781107"} Dec 03 01:49:52 crc kubenswrapper[4912]: I1203 01:49:52.156090 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" podUID="d819df61-e2e4-4960-8ff8-761a99f21850" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.250:5353: connect: connection refused" Dec 03 01:49:53 crc kubenswrapper[4912]: I1203 01:49:53.571754 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:49:53 crc kubenswrapper[4912]: E1203 01:49:53.572258 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:49:53 crc kubenswrapper[4912]: I1203 01:49:53.675401 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 01:49:53 crc kubenswrapper[4912]: I1203 01:49:53.675700 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="ceilometer-notification-agent" containerID="cri-o://1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d" gracePeriod=30 Dec 03 01:49:53 crc kubenswrapper[4912]: I1203 01:49:53.675782 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="proxy-httpd" containerID="cri-o://67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74" gracePeriod=30 Dec 03 01:49:53 crc kubenswrapper[4912]: I1203 01:49:53.675810 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="sg-core" containerID="cri-o://02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4" gracePeriod=30 Dec 03 01:49:53 crc kubenswrapper[4912]: I1203 01:49:53.675853 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="ceilometer-central-agent" containerID="cri-o://60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33" gracePeriod=30 Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.216786 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.298217 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpt78\" (UniqueName: \"kubernetes.io/projected/d819df61-e2e4-4960-8ff8-761a99f21850-kube-api-access-kpt78\") pod \"d819df61-e2e4-4960-8ff8-761a99f21850\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.298260 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-nb\") pod \"d819df61-e2e4-4960-8ff8-761a99f21850\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.298310 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-openstack-edpm-ipam\") pod \"d819df61-e2e4-4960-8ff8-761a99f21850\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.298461 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-svc\") pod \"d819df61-e2e4-4960-8ff8-761a99f21850\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.298514 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-swift-storage-0\") pod \"d819df61-e2e4-4960-8ff8-761a99f21850\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.298543 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-sb\") pod \"d819df61-e2e4-4960-8ff8-761a99f21850\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.298594 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-config\") pod \"d819df61-e2e4-4960-8ff8-761a99f21850\" (UID: \"d819df61-e2e4-4960-8ff8-761a99f21850\") " Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.311672 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d819df61-e2e4-4960-8ff8-761a99f21850-kube-api-access-kpt78" (OuterVolumeSpecName: "kube-api-access-kpt78") pod "d819df61-e2e4-4960-8ff8-761a99f21850" (UID: "d819df61-e2e4-4960-8ff8-761a99f21850"). InnerVolumeSpecName "kube-api-access-kpt78". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.381497 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "d819df61-e2e4-4960-8ff8-761a99f21850" (UID: "d819df61-e2e4-4960-8ff8-761a99f21850"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.384643 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d819df61-e2e4-4960-8ff8-761a99f21850" (UID: "d819df61-e2e4-4960-8ff8-761a99f21850"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.386991 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d819df61-e2e4-4960-8ff8-761a99f21850" (UID: "d819df61-e2e4-4960-8ff8-761a99f21850"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.394614 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d819df61-e2e4-4960-8ff8-761a99f21850" (UID: "d819df61-e2e4-4960-8ff8-761a99f21850"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.395130 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-config" (OuterVolumeSpecName: "config") pod "d819df61-e2e4-4960-8ff8-761a99f21850" (UID: "d819df61-e2e4-4960-8ff8-761a99f21850"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.401564 4912 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.401597 4912 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.401609 4912 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-config\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.401618 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpt78\" (UniqueName: \"kubernetes.io/projected/d819df61-e2e4-4960-8ff8-761a99f21850-kube-api-access-kpt78\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.401628 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.401636 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.403829 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d819df61-e2e4-4960-8ff8-761a99f21850" (UID: "d819df61-e2e4-4960-8ff8-761a99f21850"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.452382 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" event={"ID":"d819df61-e2e4-4960-8ff8-761a99f21850","Type":"ContainerDied","Data":"c22c68b7d1299076b42d44f10cf2a033f3795bc4b3221d972e171fd4ad45b47a"} Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.452444 4912 scope.go:117] "RemoveContainer" containerID="3cd1d24a28b305d242e061d5438fee63c1ba6ebc7f164d9027689fc893781107" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.452562 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.461485 4912 generic.go:334] "Generic (PLEG): container finished" podID="5fd2a312-4b62-4392-baae-306c718f24fe" containerID="60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33" exitCode=0 Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.461521 4912 generic.go:334] "Generic (PLEG): container finished" podID="5fd2a312-4b62-4392-baae-306c718f24fe" containerID="67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74" exitCode=0 Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.461532 4912 generic.go:334] "Generic (PLEG): container finished" podID="5fd2a312-4b62-4392-baae-306c718f24fe" containerID="02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4" exitCode=2 Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.461555 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerDied","Data":"60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33"} Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.461583 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerDied","Data":"67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74"} Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.461596 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerDied","Data":"02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4"} Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.487676 4912 scope.go:117] "RemoveContainer" containerID="300b2fc9e07d213a577e82624659df61c3e7a340b316c0b770cbbad5558a229c" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.491618 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf"] Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.501802 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cf7b6cbf7-ctmpf"] Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.503308 4912 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d819df61-e2e4-4960-8ff8-761a99f21850-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:54 crc kubenswrapper[4912]: I1203 01:49:54.582729 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d819df61-e2e4-4960-8ff8-761a99f21850" path="/var/lib/kubelet/pods/d819df61-e2e4-4960-8ff8-761a99f21850/volumes" Dec 03 01:49:54 crc kubenswrapper[4912]: W1203 01:49:54.902421 4912 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc963855_e556_46df_a8b0_d3a9a94bdd9e.slice/crio-5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc963855_e556_46df_a8b0_d3a9a94bdd9e.slice/crio-5c2729c3689e4b1710ebeb0c3ef41944a8fbe319e848d3935e42dcdae4c1a2ff.scope: no such file or directory Dec 03 01:49:54 crc kubenswrapper[4912]: W1203 01:49:54.903046 4912 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc963855_e556_46df_a8b0_d3a9a94bdd9e.slice/crio-conmon-15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc963855_e556_46df_a8b0_d3a9a94bdd9e.slice/crio-conmon-15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf.scope: no such file or directory Dec 03 01:49:54 crc kubenswrapper[4912]: W1203 01:49:54.903065 4912 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc963855_e556_46df_a8b0_d3a9a94bdd9e.slice/crio-15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc963855_e556_46df_a8b0_d3a9a94bdd9e.slice/crio-15f9f3a5b4e6a781dd7167684396c9e6421730ac9422caed424731184487edaf.scope: no such file or directory Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.228556 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.246:3000/\": dial tcp 10.217.0.246:3000: connect: connection refused" Dec 03 01:49:55 crc kubenswrapper[4912]: E1203 01:49:55.241749 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f9ca12c_238d_47f3_a39d_6c2df17d6baa.slice/crio-dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6880b8c_c092_43bd_9cd1_8c837be9bfe2.slice/crio-conmon-12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f9ca12c_238d_47f3_a39d_6c2df17d6baa.slice/crio-conmon-dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6880b8c_c092_43bd_9cd1_8c837be9bfe2.slice/crio-conmon-f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87.scope\": RecentStats: unable to find data in memory cache]" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.358593 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.432068 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.432155 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-config-data\") pod \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.432209 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-horizon-secret-key\") pod \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.432285 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-scripts\") pod \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.432364 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-logs\") pod \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.432472 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4tbq\" (UniqueName: \"kubernetes.io/projected/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-kube-api-access-s4tbq\") pod \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\" (UID: \"a6880b8c-c092-43bd-9cd1-8c837be9bfe2\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.432746 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-logs" (OuterVolumeSpecName: "logs") pod "a6880b8c-c092-43bd-9cd1-8c837be9bfe2" (UID: "a6880b8c-c092-43bd-9cd1-8c837be9bfe2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.433517 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-logs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.438581 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a6880b8c-c092-43bd-9cd1-8c837be9bfe2" (UID: "a6880b8c-c092-43bd-9cd1-8c837be9bfe2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.442162 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-kube-api-access-s4tbq" (OuterVolumeSpecName: "kube-api-access-s4tbq") pod "a6880b8c-c092-43bd-9cd1-8c837be9bfe2" (UID: "a6880b8c-c092-43bd-9cd1-8c837be9bfe2"). InnerVolumeSpecName "kube-api-access-s4tbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.480440 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-config-data" (OuterVolumeSpecName: "config-data") pod "a6880b8c-c092-43bd-9cd1-8c837be9bfe2" (UID: "a6880b8c-c092-43bd-9cd1-8c837be9bfe2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.481444 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-scripts" (OuterVolumeSpecName: "scripts") pod "a6880b8c-c092-43bd-9cd1-8c837be9bfe2" (UID: "a6880b8c-c092-43bd-9cd1-8c837be9bfe2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.495127 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"163bd5c4-1469-4cdd-b39f-00bf454a0adb","Type":"ContainerStarted","Data":"45ec6abc8121e46af6271619ae7d14d885ddda900269ec74f9fb4517f9f5f926"} Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.495169 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"163bd5c4-1469-4cdd-b39f-00bf454a0adb","Type":"ContainerStarted","Data":"37d0354a1991141a5162584ce83090fabc28aa97b7342591768370f5b627d42a"} Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.497298 4912 generic.go:334] "Generic (PLEG): container finished" podID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerID="dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785" exitCode=137 Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.497317 4912 generic.go:334] "Generic (PLEG): container finished" podID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerID="cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512" exitCode=137 Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.497370 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69f9d58fc9-6l6qj" event={"ID":"7f9ca12c-238d-47f3-a39d-6c2df17d6baa","Type":"ContainerDied","Data":"dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785"} Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.497390 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69f9d58fc9-6l6qj" event={"ID":"7f9ca12c-238d-47f3-a39d-6c2df17d6baa","Type":"ContainerDied","Data":"cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512"} Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.497401 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-69f9d58fc9-6l6qj" event={"ID":"7f9ca12c-238d-47f3-a39d-6c2df17d6baa","Type":"ContainerDied","Data":"cdc217ebc1f6e0d5324896094ae6191595a7c4edfa86038142ef7ed1069505d2"} Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.497418 4912 scope.go:117] "RemoveContainer" containerID="dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.497796 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-69f9d58fc9-6l6qj" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.515024 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.940830546 podStartE2EDuration="16.515008897s" podCreationTimestamp="2025-12-03 01:49:39 +0000 UTC" firstStartedPulling="2025-12-03 01:49:41.295623904 +0000 UTC m=+5166.937644464" lastFinishedPulling="2025-12-03 01:49:53.869802255 +0000 UTC m=+5179.511822815" observedRunningTime="2025-12-03 01:49:55.514613215 +0000 UTC m=+5181.156633775" watchObservedRunningTime="2025-12-03 01:49:55.515008897 +0000 UTC m=+5181.157029457" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.518885 4912 generic.go:334] "Generic (PLEG): container finished" podID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerID="12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9" exitCode=137 Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.518915 4912 generic.go:334] "Generic (PLEG): container finished" podID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerID="f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87" exitCode=137 Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.518968 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77dc7f54bf-f2rhc" event={"ID":"a6880b8c-c092-43bd-9cd1-8c837be9bfe2","Type":"ContainerDied","Data":"12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9"} Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.518995 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77dc7f54bf-f2rhc" event={"ID":"a6880b8c-c092-43bd-9cd1-8c837be9bfe2","Type":"ContainerDied","Data":"f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87"} Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.519025 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77dc7f54bf-f2rhc" event={"ID":"a6880b8c-c092-43bd-9cd1-8c837be9bfe2","Type":"ContainerDied","Data":"4f2329ffd101326d55929ee87f1a21abb689adb84d4e39785b6ea1ff6cfae440"} Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.519110 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77dc7f54bf-f2rhc" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.535245 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwv7g\" (UniqueName: \"kubernetes.io/projected/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-kube-api-access-bwv7g\") pod \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.535414 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-horizon-secret-key\") pod \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.535548 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-scripts\") pod \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.535574 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-logs\") pod \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.535633 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-config-data\") pod \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\" (UID: \"7f9ca12c-238d-47f3-a39d-6c2df17d6baa\") " Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.536198 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.536214 4912 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.536224 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.536233 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4tbq\" (UniqueName: \"kubernetes.io/projected/a6880b8c-c092-43bd-9cd1-8c837be9bfe2-kube-api-access-s4tbq\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.537587 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-logs" (OuterVolumeSpecName: "logs") pod "7f9ca12c-238d-47f3-a39d-6c2df17d6baa" (UID: "7f9ca12c-238d-47f3-a39d-6c2df17d6baa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.538998 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-kube-api-access-bwv7g" (OuterVolumeSpecName: "kube-api-access-bwv7g") pod "7f9ca12c-238d-47f3-a39d-6c2df17d6baa" (UID: "7f9ca12c-238d-47f3-a39d-6c2df17d6baa"). InnerVolumeSpecName "kube-api-access-bwv7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.551637 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "7f9ca12c-238d-47f3-a39d-6c2df17d6baa" (UID: "7f9ca12c-238d-47f3-a39d-6c2df17d6baa"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.569054 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-scripts" (OuterVolumeSpecName: "scripts") pod "7f9ca12c-238d-47f3-a39d-6c2df17d6baa" (UID: "7f9ca12c-238d-47f3-a39d-6c2df17d6baa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.572303 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-config-data" (OuterVolumeSpecName: "config-data") pod "7f9ca12c-238d-47f3-a39d-6c2df17d6baa" (UID: "7f9ca12c-238d-47f3-a39d-6c2df17d6baa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.572943 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-77dc7f54bf-f2rhc"] Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.583557 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-77dc7f54bf-f2rhc"] Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.637981 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwv7g\" (UniqueName: \"kubernetes.io/projected/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-kube-api-access-bwv7g\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.638012 4912 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.638023 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.638032 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-logs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.638040 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7f9ca12c-238d-47f3-a39d-6c2df17d6baa-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.701496 4912 scope.go:117] "RemoveContainer" containerID="cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.728004 4912 scope.go:117] "RemoveContainer" containerID="dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785" Dec 03 01:49:55 crc kubenswrapper[4912]: E1203 01:49:55.728981 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785\": container with ID starting with dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785 not found: ID does not exist" containerID="dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.729023 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785"} err="failed to get container status \"dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785\": rpc error: code = NotFound desc = could not find container \"dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785\": container with ID starting with dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785 not found: ID does not exist" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.729051 4912 scope.go:117] "RemoveContainer" containerID="cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512" Dec 03 01:49:55 crc kubenswrapper[4912]: E1203 01:49:55.729383 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512\": container with ID starting with cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512 not found: ID does not exist" containerID="cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.729422 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512"} err="failed to get container status \"cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512\": rpc error: code = NotFound desc = could not find container \"cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512\": container with ID starting with cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512 not found: ID does not exist" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.729463 4912 scope.go:117] "RemoveContainer" containerID="dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.729696 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785"} err="failed to get container status \"dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785\": rpc error: code = NotFound desc = could not find container \"dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785\": container with ID starting with dc7ca9101f89ce9882f8a56b87e9c8ad3d9e8ddab500be8ffc3188e5d022c785 not found: ID does not exist" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.729717 4912 scope.go:117] "RemoveContainer" containerID="cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.732825 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512"} err="failed to get container status \"cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512\": rpc error: code = NotFound desc = could not find container \"cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512\": container with ID starting with cfaa19635de4b38f342295235ff02cad554a1935845ac6549986fcdc9286d512 not found: ID does not exist" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.732853 4912 scope.go:117] "RemoveContainer" containerID="12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9" Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.934808 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-69f9d58fc9-6l6qj"] Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.944231 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-69f9d58fc9-6l6qj"] Dec 03 01:49:55 crc kubenswrapper[4912]: I1203 01:49:55.973946 4912 scope.go:117] "RemoveContainer" containerID="f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.101112 4912 scope.go:117] "RemoveContainer" containerID="12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9" Dec 03 01:49:56 crc kubenswrapper[4912]: E1203 01:49:56.112061 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9\": container with ID starting with 12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9 not found: ID does not exist" containerID="12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.112114 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9"} err="failed to get container status \"12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9\": rpc error: code = NotFound desc = could not find container \"12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9\": container with ID starting with 12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9 not found: ID does not exist" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.112149 4912 scope.go:117] "RemoveContainer" containerID="f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87" Dec 03 01:49:56 crc kubenswrapper[4912]: E1203 01:49:56.112651 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87\": container with ID starting with f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87 not found: ID does not exist" containerID="f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.112684 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87"} err="failed to get container status \"f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87\": rpc error: code = NotFound desc = could not find container \"f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87\": container with ID starting with f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87 not found: ID does not exist" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.112702 4912 scope.go:117] "RemoveContainer" containerID="12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.112952 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9"} err="failed to get container status \"12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9\": rpc error: code = NotFound desc = could not find container \"12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9\": container with ID starting with 12cd262e59cd657c3c01d2fa00b1d70b1c26f07cc0a950e6b3090fd4cffbf5c9 not found: ID does not exist" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.112977 4912 scope.go:117] "RemoveContainer" containerID="f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.113300 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87"} err="failed to get container status \"f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87\": rpc error: code = NotFound desc = could not find container \"f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87\": container with ID starting with f834725e6945e85a5aca7953c5609442e6d2f6970e997bc2abad02c65437bc87 not found: ID does not exist" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.608134 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" path="/var/lib/kubelet/pods/7f9ca12c-238d-47f3-a39d-6c2df17d6baa/volumes" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.609168 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" path="/var/lib/kubelet/pods/a6880b8c-c092-43bd-9cd1-8c837be9bfe2/volumes" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.612688 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.624821 4912 generic.go:334] "Generic (PLEG): container finished" podID="5fd2a312-4b62-4392-baae-306c718f24fe" containerID="1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d" exitCode=0 Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.625828 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerDied","Data":"1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d"} Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.625860 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fd2a312-4b62-4392-baae-306c718f24fe","Type":"ContainerDied","Data":"ccefc08d02a51f3a44a53a1937d41afaae3056296ab10bcea1675d228a920287"} Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.625887 4912 scope.go:117] "RemoveContainer" containerID="60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.653502 4912 scope.go:117] "RemoveContainer" containerID="67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.690153 4912 scope.go:117] "RemoveContainer" containerID="02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.691385 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-run-httpd\") pod \"5fd2a312-4b62-4392-baae-306c718f24fe\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.691517 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-sg-core-conf-yaml\") pod \"5fd2a312-4b62-4392-baae-306c718f24fe\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.691576 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-scripts\") pod \"5fd2a312-4b62-4392-baae-306c718f24fe\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.691604 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-config-data\") pod \"5fd2a312-4b62-4392-baae-306c718f24fe\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.691634 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-combined-ca-bundle\") pod \"5fd2a312-4b62-4392-baae-306c718f24fe\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.691827 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-ceilometer-tls-certs\") pod \"5fd2a312-4b62-4392-baae-306c718f24fe\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.691861 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzsz2\" (UniqueName: \"kubernetes.io/projected/5fd2a312-4b62-4392-baae-306c718f24fe-kube-api-access-nzsz2\") pod \"5fd2a312-4b62-4392-baae-306c718f24fe\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.691987 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-log-httpd\") pod \"5fd2a312-4b62-4392-baae-306c718f24fe\" (UID: \"5fd2a312-4b62-4392-baae-306c718f24fe\") " Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.701120 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5fd2a312-4b62-4392-baae-306c718f24fe" (UID: "5fd2a312-4b62-4392-baae-306c718f24fe"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.704236 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5fd2a312-4b62-4392-baae-306c718f24fe" (UID: "5fd2a312-4b62-4392-baae-306c718f24fe"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.704693 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-scripts" (OuterVolumeSpecName: "scripts") pod "5fd2a312-4b62-4392-baae-306c718f24fe" (UID: "5fd2a312-4b62-4392-baae-306c718f24fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.711004 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fd2a312-4b62-4392-baae-306c718f24fe-kube-api-access-nzsz2" (OuterVolumeSpecName: "kube-api-access-nzsz2") pod "5fd2a312-4b62-4392-baae-306c718f24fe" (UID: "5fd2a312-4b62-4392-baae-306c718f24fe"). InnerVolumeSpecName "kube-api-access-nzsz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.757623 4912 scope.go:117] "RemoveContainer" containerID="1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.782223 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5fd2a312-4b62-4392-baae-306c718f24fe" (UID: "5fd2a312-4b62-4392-baae-306c718f24fe"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.808628 4912 scope.go:117] "RemoveContainer" containerID="60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33" Dec 03 01:49:56 crc kubenswrapper[4912]: E1203 01:49:56.809364 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33\": container with ID starting with 60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33 not found: ID does not exist" containerID="60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.809403 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33"} err="failed to get container status \"60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33\": rpc error: code = NotFound desc = could not find container \"60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33\": container with ID starting with 60ede55e2dea0d7fa4c3716f65192615a209651fa89563ee54dbaf30a14e6f33 not found: ID does not exist" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.809445 4912 scope.go:117] "RemoveContainer" containerID="67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.811175 4912 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.811192 4912 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fd2a312-4b62-4392-baae-306c718f24fe-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.811202 4912 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.811210 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.811219 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzsz2\" (UniqueName: \"kubernetes.io/projected/5fd2a312-4b62-4392-baae-306c718f24fe-kube-api-access-nzsz2\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:56 crc kubenswrapper[4912]: E1203 01:49:56.813598 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74\": container with ID starting with 67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74 not found: ID does not exist" containerID="67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.813649 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74"} err="failed to get container status \"67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74\": rpc error: code = NotFound desc = could not find container \"67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74\": container with ID starting with 67dbadfc32ffc434237dff48c7614b6545d2a5bbc9bfb92d7649d348b3be6e74 not found: ID does not exist" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.813677 4912 scope.go:117] "RemoveContainer" containerID="02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4" Dec 03 01:49:56 crc kubenswrapper[4912]: E1203 01:49:56.814366 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4\": container with ID starting with 02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4 not found: ID does not exist" containerID="02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.814387 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4"} err="failed to get container status \"02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4\": rpc error: code = NotFound desc = could not find container \"02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4\": container with ID starting with 02c55f33e51a2b4b49c6702232d1c86c1fd12672e9009f6eec58657ce8e6f3c4 not found: ID does not exist" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.814404 4912 scope.go:117] "RemoveContainer" containerID="1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d" Dec 03 01:49:56 crc kubenswrapper[4912]: E1203 01:49:56.818622 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d\": container with ID starting with 1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d not found: ID does not exist" containerID="1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.818643 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "5fd2a312-4b62-4392-baae-306c718f24fe" (UID: "5fd2a312-4b62-4392-baae-306c718f24fe"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.818669 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d"} err="failed to get container status \"1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d\": rpc error: code = NotFound desc = could not find container \"1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d\": container with ID starting with 1b9d84cc5d9483719f4ee71dcdc2d84f2ad3a723d68d2c50198a2b5421467a1d not found: ID does not exist" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.914499 4912 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.939653 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fd2a312-4b62-4392-baae-306c718f24fe" (UID: "5fd2a312-4b62-4392-baae-306c718f24fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:56 crc kubenswrapper[4912]: I1203 01:49:56.961682 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-config-data" (OuterVolumeSpecName: "config-data") pod "5fd2a312-4b62-4392-baae-306c718f24fe" (UID: "5fd2a312-4b62-4392-baae-306c718f24fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.016465 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.016494 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fd2a312-4b62-4392-baae-306c718f24fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.635702 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.681155 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.698283 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710179 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710701 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="proxy-httpd" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710720 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="proxy-httpd" Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710739 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d819df61-e2e4-4960-8ff8-761a99f21850" containerName="dnsmasq-dns" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710745 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d819df61-e2e4-4960-8ff8-761a99f21850" containerName="dnsmasq-dns" Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710763 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerName="horizon-log" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710769 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerName="horizon-log" Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710781 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="ceilometer-notification-agent" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710788 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="ceilometer-notification-agent" Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710804 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerName="horizon-log" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710811 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerName="horizon-log" Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710825 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerName="horizon" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710831 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerName="horizon" Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710845 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="ceilometer-central-agent" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710851 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="ceilometer-central-agent" Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710861 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerName="horizon" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710867 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerName="horizon" Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710886 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="sg-core" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710891 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="sg-core" Dec 03 01:49:57 crc kubenswrapper[4912]: E1203 01:49:57.710900 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d819df61-e2e4-4960-8ff8-761a99f21850" containerName="init" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.710905 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d819df61-e2e4-4960-8ff8-761a99f21850" containerName="init" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.711097 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="ceilometer-notification-agent" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.711110 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="sg-core" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.711121 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d819df61-e2e4-4960-8ff8-761a99f21850" containerName="dnsmasq-dns" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.711129 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="ceilometer-central-agent" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.711145 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerName="horizon" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.711156 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" containerName="proxy-httpd" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.711174 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f9ca12c-238d-47f3-a39d-6c2df17d6baa" containerName="horizon-log" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.711185 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerName="horizon-log" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.711195 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6880b8c-c092-43bd-9cd1-8c837be9bfe2" containerName="horizon" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.713158 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.721117 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.721323 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.721460 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.726124 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.832839 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.832943 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.833006 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbc8b9dd-35e3-4d6b-8df6-e44368688196-log-httpd\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.833202 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbc8b9dd-35e3-4d6b-8df6-e44368688196-run-httpd\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.833424 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.833526 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-config-data\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.833621 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-scripts\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.833671 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncfg7\" (UniqueName: \"kubernetes.io/projected/bbc8b9dd-35e3-4d6b-8df6-e44368688196-kube-api-access-ncfg7\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.935518 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.935594 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.935711 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbc8b9dd-35e3-4d6b-8df6-e44368688196-log-httpd\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.936256 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbc8b9dd-35e3-4d6b-8df6-e44368688196-log-httpd\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.936657 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbc8b9dd-35e3-4d6b-8df6-e44368688196-run-httpd\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.936697 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbc8b9dd-35e3-4d6b-8df6-e44368688196-run-httpd\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.936836 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.937288 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-config-data\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.937365 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-scripts\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.937401 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncfg7\" (UniqueName: \"kubernetes.io/projected/bbc8b9dd-35e3-4d6b-8df6-e44368688196-kube-api-access-ncfg7\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.941300 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.941651 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.943912 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-scripts\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.944534 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-config-data\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.952983 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbc8b9dd-35e3-4d6b-8df6-e44368688196-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:57 crc kubenswrapper[4912]: I1203 01:49:57.975446 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncfg7\" (UniqueName: \"kubernetes.io/projected/bbc8b9dd-35e3-4d6b-8df6-e44368688196-kube-api-access-ncfg7\") pod \"ceilometer-0\" (UID: \"bbc8b9dd-35e3-4d6b-8df6-e44368688196\") " pod="openstack/ceilometer-0" Dec 03 01:49:58 crc kubenswrapper[4912]: I1203 01:49:58.044567 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 01:49:58 crc kubenswrapper[4912]: I1203 01:49:58.553526 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 01:49:58 crc kubenswrapper[4912]: I1203 01:49:58.600038 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fd2a312-4b62-4392-baae-306c718f24fe" path="/var/lib/kubelet/pods/5fd2a312-4b62-4392-baae-306c718f24fe/volumes" Dec 03 01:49:58 crc kubenswrapper[4912]: I1203 01:49:58.653946 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbc8b9dd-35e3-4d6b-8df6-e44368688196","Type":"ContainerStarted","Data":"e5a9e50a034366da14077ed7aaa9d3dfd3a4c513b9a75ca4ffc964854a08fcb5"} Dec 03 01:49:58 crc kubenswrapper[4912]: I1203 01:49:58.905616 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:49:58 crc kubenswrapper[4912]: I1203 01:49:58.911555 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:49:59 crc kubenswrapper[4912]: I1203 01:49:59.674037 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbc8b9dd-35e3-4d6b-8df6-e44368688196","Type":"ContainerStarted","Data":"f99f064fe16c278ae4fda95dc3d3d7f14e7100035b310868eea0d9c722e6ad82"} Dec 03 01:49:59 crc kubenswrapper[4912]: I1203 01:49:59.898958 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 03 01:50:00 crc kubenswrapper[4912]: I1203 01:50:00.690105 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbc8b9dd-35e3-4d6b-8df6-e44368688196","Type":"ContainerStarted","Data":"4a143dda9ed377ec6a996bcb1bbcae1aae5f88868a21a99ecdd9c3bee752b711"} Dec 03 01:50:00 crc kubenswrapper[4912]: I1203 01:50:00.779005 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:50:00 crc kubenswrapper[4912]: I1203 01:50:00.948222 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6d4dbbf8d-5vgmn" Dec 03 01:50:01 crc kubenswrapper[4912]: I1203 01:50:01.010228 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d9dc8d44b-7tljg"] Dec 03 01:50:01 crc kubenswrapper[4912]: I1203 01:50:01.612033 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 03 01:50:01 crc kubenswrapper[4912]: I1203 01:50:01.661920 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 01:50:01 crc kubenswrapper[4912]: I1203 01:50:01.700378 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbc8b9dd-35e3-4d6b-8df6-e44368688196","Type":"ContainerStarted","Data":"ca7a7f710051f405e4cd0ae355c5a623fcd546f855a567373b59624884e1a184"} Dec 03 01:50:01 crc kubenswrapper[4912]: I1203 01:50:01.700694 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerName="manila-scheduler" containerID="cri-o://6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3" gracePeriod=30 Dec 03 01:50:01 crc kubenswrapper[4912]: I1203 01:50:01.700817 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerName="probe" containerID="cri-o://8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14" gracePeriod=30 Dec 03 01:50:01 crc kubenswrapper[4912]: I1203 01:50:01.701871 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d9dc8d44b-7tljg" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon" containerID="cri-o://d6aa5aaf08c7d099e3be624e725ccb79ec7adb51678136411bc203ceed6842ea" gracePeriod=30 Dec 03 01:50:01 crc kubenswrapper[4912]: I1203 01:50:01.701832 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d9dc8d44b-7tljg" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon-log" containerID="cri-o://c77922fad380b1690cfe80e8b1b9cb7b6921b63e8a0da1cdc0b8624aac711282" gracePeriod=30 Dec 03 01:50:02 crc kubenswrapper[4912]: I1203 01:50:02.718383 4912 generic.go:334] "Generic (PLEG): container finished" podID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerID="8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14" exitCode=0 Dec 03 01:50:02 crc kubenswrapper[4912]: I1203 01:50:02.718579 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1","Type":"ContainerDied","Data":"8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14"} Dec 03 01:50:03 crc kubenswrapper[4912]: I1203 01:50:03.735480 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbc8b9dd-35e3-4d6b-8df6-e44368688196","Type":"ContainerStarted","Data":"2670b1c1795a6e13d4a68363c354976b658d3eb4d8d669911f9ab6212576e811"} Dec 03 01:50:03 crc kubenswrapper[4912]: I1203 01:50:03.736738 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 01:50:03 crc kubenswrapper[4912]: I1203 01:50:03.764036 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.848462964 podStartE2EDuration="6.764015171s" podCreationTimestamp="2025-12-03 01:49:57 +0000 UTC" firstStartedPulling="2025-12-03 01:49:58.547723285 +0000 UTC m=+5184.189743835" lastFinishedPulling="2025-12-03 01:50:02.463275482 +0000 UTC m=+5188.105296042" observedRunningTime="2025-12-03 01:50:03.7546785 +0000 UTC m=+5189.396699060" watchObservedRunningTime="2025-12-03 01:50:03.764015171 +0000 UTC m=+5189.406035731" Dec 03 01:50:05 crc kubenswrapper[4912]: I1203 01:50:05.757115 4912 generic.go:334] "Generic (PLEG): container finished" podID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerID="d6aa5aaf08c7d099e3be624e725ccb79ec7adb51678136411bc203ceed6842ea" exitCode=0 Dec 03 01:50:05 crc kubenswrapper[4912]: I1203 01:50:05.757178 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d9dc8d44b-7tljg" event={"ID":"08762529-e1c5-4358-bc43-c0d81a4c6df6","Type":"ContainerDied","Data":"d6aa5aaf08c7d099e3be624e725ccb79ec7adb51678136411bc203ceed6842ea"} Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.339694 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d9dc8d44b-7tljg" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.69:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.69:8443: connect: connection refused" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.412189 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.569941 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data-custom\") pod \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.570114 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data\") pod \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.571067 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-scripts\") pod \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.571152 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-etc-machine-id\") pod \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.571203 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6xsz\" (UniqueName: \"kubernetes.io/projected/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-kube-api-access-v6xsz\") pod \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.571246 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-combined-ca-bundle\") pod \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\" (UID: \"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1\") " Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.571271 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" (UID: "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.572206 4912 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.579540 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-scripts" (OuterVolumeSpecName: "scripts") pod "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" (UID: "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.589580 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-kube-api-access-v6xsz" (OuterVolumeSpecName: "kube-api-access-v6xsz") pod "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" (UID: "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1"). InnerVolumeSpecName "kube-api-access-v6xsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.590607 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" (UID: "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.674892 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.675314 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6xsz\" (UniqueName: \"kubernetes.io/projected/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-kube-api-access-v6xsz\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.675461 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.686646 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" (UID: "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.765156 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data" (OuterVolumeSpecName: "config-data") pod "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" (UID: "509cb440-9d5a-4996-a0a6-f6bed4f1fdf1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.772505 4912 generic.go:334] "Generic (PLEG): container finished" podID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerID="6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3" exitCode=0 Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.772543 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1","Type":"ContainerDied","Data":"6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3"} Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.772568 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"509cb440-9d5a-4996-a0a6-f6bed4f1fdf1","Type":"ContainerDied","Data":"5142ee9482a19b7e3791df7ee0548a2293eed18123eba8baec80ac00a735bdd0"} Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.772585 4912 scope.go:117] "RemoveContainer" containerID="8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.772708 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.778326 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.778365 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.818751 4912 scope.go:117] "RemoveContainer" containerID="6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.821587 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.840656 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.864663 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 01:50:06 crc kubenswrapper[4912]: E1203 01:50:06.865289 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerName="manila-scheduler" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.865312 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerName="manila-scheduler" Dec 03 01:50:06 crc kubenswrapper[4912]: E1203 01:50:06.865328 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerName="probe" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.865336 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerName="probe" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.865614 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerName="manila-scheduler" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.865638 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" containerName="probe" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.879251 4912 scope.go:117] "RemoveContainer" containerID="8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.884227 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 01:50:06 crc kubenswrapper[4912]: E1203 01:50:06.884301 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14\": container with ID starting with 8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14 not found: ID does not exist" containerID="8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.884337 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14"} err="failed to get container status \"8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14\": rpc error: code = NotFound desc = could not find container \"8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14\": container with ID starting with 8a4ea2bfaa043294ffe6f253666589d37dbf6c1d1034378a53491adf09c28c14 not found: ID does not exist" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.884361 4912 scope.go:117] "RemoveContainer" containerID="6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3" Dec 03 01:50:06 crc kubenswrapper[4912]: E1203 01:50:06.886555 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3\": container with ID starting with 6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3 not found: ID does not exist" containerID="6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.886605 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3"} err="failed to get container status \"6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3\": rpc error: code = NotFound desc = could not find container \"6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3\": container with ID starting with 6b0605a3bec37da325786fac9f00df9bdfc8d4f482494942f3e476c859c7f5f3 not found: ID does not exist" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.886652 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.903951 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.981949 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psxp5\" (UniqueName: \"kubernetes.io/projected/dc97be40-fb7b-4af2-af77-45f0f360c8cb-kube-api-access-psxp5\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.982030 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-config-data\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.982055 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.982382 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-scripts\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.982750 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:06 crc kubenswrapper[4912]: I1203 01:50:06.982951 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc97be40-fb7b-4af2-af77-45f0f360c8cb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.085474 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-config-data\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.085530 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.085583 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-scripts\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.085667 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.085736 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc97be40-fb7b-4af2-af77-45f0f360c8cb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.085903 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc97be40-fb7b-4af2-af77-45f0f360c8cb-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.085928 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psxp5\" (UniqueName: \"kubernetes.io/projected/dc97be40-fb7b-4af2-af77-45f0f360c8cb-kube-api-access-psxp5\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.089541 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-scripts\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.090405 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.092513 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-config-data\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.093419 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc97be40-fb7b-4af2-af77-45f0f360c8cb-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.110365 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psxp5\" (UniqueName: \"kubernetes.io/projected/dc97be40-fb7b-4af2-af77-45f0f360c8cb-kube-api-access-psxp5\") pod \"manila-scheduler-0\" (UID: \"dc97be40-fb7b-4af2-af77-45f0f360c8cb\") " pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.217053 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.618716 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:50:07 crc kubenswrapper[4912]: E1203 01:50:07.619471 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:50:07 crc kubenswrapper[4912]: I1203 01:50:07.986181 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 01:50:08 crc kubenswrapper[4912]: I1203 01:50:08.259743 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 03 01:50:08 crc kubenswrapper[4912]: I1203 01:50:08.585713 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="509cb440-9d5a-4996-a0a6-f6bed4f1fdf1" path="/var/lib/kubelet/pods/509cb440-9d5a-4996-a0a6-f6bed4f1fdf1/volumes" Dec 03 01:50:08 crc kubenswrapper[4912]: I1203 01:50:08.796243 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"dc97be40-fb7b-4af2-af77-45f0f360c8cb","Type":"ContainerStarted","Data":"bf577a7fa732fda3fc2df498be4ee34e67771ccd1a2e0412b05ac5bb856e3c98"} Dec 03 01:50:08 crc kubenswrapper[4912]: I1203 01:50:08.796541 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"dc97be40-fb7b-4af2-af77-45f0f360c8cb","Type":"ContainerStarted","Data":"890071e0655d6e3ba785cf8acc69ff1e083883a386837e4e20d8d63f1ac514c9"} Dec 03 01:50:09 crc kubenswrapper[4912]: I1203 01:50:09.810680 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"dc97be40-fb7b-4af2-af77-45f0f360c8cb","Type":"ContainerStarted","Data":"24677dcb3a3de3fad403c08315037886e8778b07e82eeb7bf96f7a95074a0749"} Dec 03 01:50:09 crc kubenswrapper[4912]: I1203 01:50:09.842102 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.8420746489999997 podStartE2EDuration="3.842074649s" podCreationTimestamp="2025-12-03 01:50:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:50:09.835670137 +0000 UTC m=+5195.477690777" watchObservedRunningTime="2025-12-03 01:50:09.842074649 +0000 UTC m=+5195.484095249" Dec 03 01:50:11 crc kubenswrapper[4912]: I1203 01:50:11.471842 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 03 01:50:11 crc kubenswrapper[4912]: I1203 01:50:11.549230 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 01:50:11 crc kubenswrapper[4912]: I1203 01:50:11.835188 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerName="manila-share" containerID="cri-o://37d0354a1991141a5162584ce83090fabc28aa97b7342591768370f5b627d42a" gracePeriod=30 Dec 03 01:50:11 crc kubenswrapper[4912]: I1203 01:50:11.835716 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerName="probe" containerID="cri-o://45ec6abc8121e46af6271619ae7d14d885ddda900269ec74f9fb4517f9f5f926" gracePeriod=30 Dec 03 01:50:12 crc kubenswrapper[4912]: I1203 01:50:12.845233 4912 generic.go:334] "Generic (PLEG): container finished" podID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerID="45ec6abc8121e46af6271619ae7d14d885ddda900269ec74f9fb4517f9f5f926" exitCode=0 Dec 03 01:50:12 crc kubenswrapper[4912]: I1203 01:50:12.845502 4912 generic.go:334] "Generic (PLEG): container finished" podID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerID="37d0354a1991141a5162584ce83090fabc28aa97b7342591768370f5b627d42a" exitCode=1 Dec 03 01:50:12 crc kubenswrapper[4912]: I1203 01:50:12.845302 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"163bd5c4-1469-4cdd-b39f-00bf454a0adb","Type":"ContainerDied","Data":"45ec6abc8121e46af6271619ae7d14d885ddda900269ec74f9fb4517f9f5f926"} Dec 03 01:50:12 crc kubenswrapper[4912]: I1203 01:50:12.845538 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"163bd5c4-1469-4cdd-b39f-00bf454a0adb","Type":"ContainerDied","Data":"37d0354a1991141a5162584ce83090fabc28aa97b7342591768370f5b627d42a"} Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.431520 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.576046 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-ceph\") pod \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.576121 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data\") pod \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.576172 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-etc-machine-id\") pod \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.576238 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-combined-ca-bundle\") pod \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.576292 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmv4f\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-kube-api-access-nmv4f\") pod \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.576331 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-scripts\") pod \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.576446 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-var-lib-manila\") pod \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.576477 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data-custom\") pod \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\" (UID: \"163bd5c4-1469-4cdd-b39f-00bf454a0adb\") " Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.577038 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "163bd5c4-1469-4cdd-b39f-00bf454a0adb" (UID: "163bd5c4-1469-4cdd-b39f-00bf454a0adb"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.577105 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "163bd5c4-1469-4cdd-b39f-00bf454a0adb" (UID: "163bd5c4-1469-4cdd-b39f-00bf454a0adb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.584811 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "163bd5c4-1469-4cdd-b39f-00bf454a0adb" (UID: "163bd5c4-1469-4cdd-b39f-00bf454a0adb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.584963 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-kube-api-access-nmv4f" (OuterVolumeSpecName: "kube-api-access-nmv4f") pod "163bd5c4-1469-4cdd-b39f-00bf454a0adb" (UID: "163bd5c4-1469-4cdd-b39f-00bf454a0adb"). InnerVolumeSpecName "kube-api-access-nmv4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.586042 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-ceph" (OuterVolumeSpecName: "ceph") pod "163bd5c4-1469-4cdd-b39f-00bf454a0adb" (UID: "163bd5c4-1469-4cdd-b39f-00bf454a0adb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.587123 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-scripts" (OuterVolumeSpecName: "scripts") pod "163bd5c4-1469-4cdd-b39f-00bf454a0adb" (UID: "163bd5c4-1469-4cdd-b39f-00bf454a0adb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.678854 4912 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.679036 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmv4f\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-kube-api-access-nmv4f\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.679109 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.679185 4912 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/163bd5c4-1469-4cdd-b39f-00bf454a0adb-var-lib-manila\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.679239 4912 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.679300 4912 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/163bd5c4-1469-4cdd-b39f-00bf454a0adb-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.680071 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "163bd5c4-1469-4cdd-b39f-00bf454a0adb" (UID: "163bd5c4-1469-4cdd-b39f-00bf454a0adb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.708654 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data" (OuterVolumeSpecName: "config-data") pod "163bd5c4-1469-4cdd-b39f-00bf454a0adb" (UID: "163bd5c4-1469-4cdd-b39f-00bf454a0adb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.781629 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.781666 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163bd5c4-1469-4cdd-b39f-00bf454a0adb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.854559 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"163bd5c4-1469-4cdd-b39f-00bf454a0adb","Type":"ContainerDied","Data":"2eee6918fa22c516383e9e6eb3361b197df8e7db124b3e659598cf7f331398fd"} Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.854606 4912 scope.go:117] "RemoveContainer" containerID="45ec6abc8121e46af6271619ae7d14d885ddda900269ec74f9fb4517f9f5f926" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.854641 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.889673 4912 scope.go:117] "RemoveContainer" containerID="37d0354a1991141a5162584ce83090fabc28aa97b7342591768370f5b627d42a" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.890462 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.900978 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.926128 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 01:50:13 crc kubenswrapper[4912]: E1203 01:50:13.926627 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerName="probe" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.926648 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerName="probe" Dec 03 01:50:13 crc kubenswrapper[4912]: E1203 01:50:13.926719 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerName="manila-share" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.926729 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerName="manila-share" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.926990 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerName="probe" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.927048 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" containerName="manila-share" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.928477 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.930723 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.954419 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.986895 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-config-data\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.987159 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-scripts\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.987247 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.987396 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b2a55784-64e0-4397-a975-c0b069cbd08f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.987557 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b2a55784-64e0-4397-a975-c0b069cbd08f-ceph\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.987587 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b2a55784-64e0-4397-a975-c0b069cbd08f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.987621 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvq9k\" (UniqueName: \"kubernetes.io/projected/b2a55784-64e0-4397-a975-c0b069cbd08f-kube-api-access-wvq9k\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:13 crc kubenswrapper[4912]: I1203 01:50:13.987652 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.089699 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b2a55784-64e0-4397-a975-c0b069cbd08f-ceph\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.089780 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b2a55784-64e0-4397-a975-c0b069cbd08f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.089849 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvq9k\" (UniqueName: \"kubernetes.io/projected/b2a55784-64e0-4397-a975-c0b069cbd08f-kube-api-access-wvq9k\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.089911 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.089999 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-config-data\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.090090 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-scripts\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.090142 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.090256 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b2a55784-64e0-4397-a975-c0b069cbd08f-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.090274 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b2a55784-64e0-4397-a975-c0b069cbd08f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.090336 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b2a55784-64e0-4397-a975-c0b069cbd08f-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.094596 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b2a55784-64e0-4397-a975-c0b069cbd08f-ceph\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.095116 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.096259 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-config-data\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.099077 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.100578 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2a55784-64e0-4397-a975-c0b069cbd08f-scripts\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.113045 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvq9k\" (UniqueName: \"kubernetes.io/projected/b2a55784-64e0-4397-a975-c0b069cbd08f-kube-api-access-wvq9k\") pod \"manila-share-share1-0\" (UID: \"b2a55784-64e0-4397-a975-c0b069cbd08f\") " pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.244872 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.584841 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="163bd5c4-1469-4cdd-b39f-00bf454a0adb" path="/var/lib/kubelet/pods/163bd5c4-1469-4cdd-b39f-00bf454a0adb/volumes" Dec 03 01:50:14 crc kubenswrapper[4912]: I1203 01:50:14.842747 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 01:50:14 crc kubenswrapper[4912]: W1203 01:50:14.864062 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2a55784_64e0_4397_a975_c0b069cbd08f.slice/crio-d2f13483b00dc45a5e7336e2d3640a907256dc5f6bba9db68e44304d6028b90e WatchSource:0}: Error finding container d2f13483b00dc45a5e7336e2d3640a907256dc5f6bba9db68e44304d6028b90e: Status 404 returned error can't find the container with id d2f13483b00dc45a5e7336e2d3640a907256dc5f6bba9db68e44304d6028b90e Dec 03 01:50:15 crc kubenswrapper[4912]: I1203 01:50:15.879773 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b2a55784-64e0-4397-a975-c0b069cbd08f","Type":"ContainerStarted","Data":"b0fefb24dd8e2bda5e82bfb70524ce43dd686ac5bd7070f972df10831a0fdd9c"} Dec 03 01:50:15 crc kubenswrapper[4912]: I1203 01:50:15.880058 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b2a55784-64e0-4397-a975-c0b069cbd08f","Type":"ContainerStarted","Data":"4ad6c03e07fb7a3a07bd34e17fe64c09eee72fa49eaebe7c601a1f4e29ef5622"} Dec 03 01:50:15 crc kubenswrapper[4912]: I1203 01:50:15.880069 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b2a55784-64e0-4397-a975-c0b069cbd08f","Type":"ContainerStarted","Data":"d2f13483b00dc45a5e7336e2d3640a907256dc5f6bba9db68e44304d6028b90e"} Dec 03 01:50:15 crc kubenswrapper[4912]: I1203 01:50:15.923693 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.9236740919999997 podStartE2EDuration="2.923674092s" podCreationTimestamp="2025-12-03 01:50:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 01:50:15.899279175 +0000 UTC m=+5201.541299735" watchObservedRunningTime="2025-12-03 01:50:15.923674092 +0000 UTC m=+5201.565694652" Dec 03 01:50:16 crc kubenswrapper[4912]: I1203 01:50:16.340471 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d9dc8d44b-7tljg" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.69:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.69:8443: connect: connection refused" Dec 03 01:50:17 crc kubenswrapper[4912]: I1203 01:50:17.218290 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 03 01:50:20 crc kubenswrapper[4912]: I1203 01:50:20.572060 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:50:20 crc kubenswrapper[4912]: E1203 01:50:20.573615 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:50:24 crc kubenswrapper[4912]: I1203 01:50:24.245408 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 03 01:50:26 crc kubenswrapper[4912]: I1203 01:50:26.339881 4912 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d9dc8d44b-7tljg" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.69:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.69:8443: connect: connection refused" Dec 03 01:50:26 crc kubenswrapper[4912]: I1203 01:50:26.340682 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:50:28 crc kubenswrapper[4912]: I1203 01:50:28.057769 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 01:50:28 crc kubenswrapper[4912]: I1203 01:50:28.987693 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 03 01:50:32 crc kubenswrapper[4912]: I1203 01:50:32.118679 4912 generic.go:334] "Generic (PLEG): container finished" podID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerID="c77922fad380b1690cfe80e8b1b9cb7b6921b63e8a0da1cdc0b8624aac711282" exitCode=137 Dec 03 01:50:32 crc kubenswrapper[4912]: I1203 01:50:32.118780 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d9dc8d44b-7tljg" event={"ID":"08762529-e1c5-4358-bc43-c0d81a4c6df6","Type":"ContainerDied","Data":"c77922fad380b1690cfe80e8b1b9cb7b6921b63e8a0da1cdc0b8624aac711282"} Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.036296 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.131889 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d9dc8d44b-7tljg" event={"ID":"08762529-e1c5-4358-bc43-c0d81a4c6df6","Type":"ContainerDied","Data":"7a1396ea6289d0dbfc574b976e06d26a605d4e6fefca772b5a1356af744523f2"} Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.131955 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d9dc8d44b-7tljg" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.132256 4912 scope.go:117] "RemoveContainer" containerID="d6aa5aaf08c7d099e3be624e725ccb79ec7adb51678136411bc203ceed6842ea" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.139579 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08762529-e1c5-4358-bc43-c0d81a4c6df6-logs\") pod \"08762529-e1c5-4358-bc43-c0d81a4c6df6\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.139627 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-secret-key\") pod \"08762529-e1c5-4358-bc43-c0d81a4c6df6\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.139831 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-combined-ca-bundle\") pod \"08762529-e1c5-4358-bc43-c0d81a4c6df6\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.139886 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-config-data\") pod \"08762529-e1c5-4358-bc43-c0d81a4c6df6\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.140030 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-tls-certs\") pod \"08762529-e1c5-4358-bc43-c0d81a4c6df6\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.140074 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcxrz\" (UniqueName: \"kubernetes.io/projected/08762529-e1c5-4358-bc43-c0d81a4c6df6-kube-api-access-tcxrz\") pod \"08762529-e1c5-4358-bc43-c0d81a4c6df6\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.140097 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-scripts\") pod \"08762529-e1c5-4358-bc43-c0d81a4c6df6\" (UID: \"08762529-e1c5-4358-bc43-c0d81a4c6df6\") " Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.140153 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08762529-e1c5-4358-bc43-c0d81a4c6df6-logs" (OuterVolumeSpecName: "logs") pod "08762529-e1c5-4358-bc43-c0d81a4c6df6" (UID: "08762529-e1c5-4358-bc43-c0d81a4c6df6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.140660 4912 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08762529-e1c5-4358-bc43-c0d81a4c6df6-logs\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.148337 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08762529-e1c5-4358-bc43-c0d81a4c6df6-kube-api-access-tcxrz" (OuterVolumeSpecName: "kube-api-access-tcxrz") pod "08762529-e1c5-4358-bc43-c0d81a4c6df6" (UID: "08762529-e1c5-4358-bc43-c0d81a4c6df6"). InnerVolumeSpecName "kube-api-access-tcxrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.164303 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "08762529-e1c5-4358-bc43-c0d81a4c6df6" (UID: "08762529-e1c5-4358-bc43-c0d81a4c6df6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.175744 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-scripts" (OuterVolumeSpecName: "scripts") pod "08762529-e1c5-4358-bc43-c0d81a4c6df6" (UID: "08762529-e1c5-4358-bc43-c0d81a4c6df6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.176781 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-config-data" (OuterVolumeSpecName: "config-data") pod "08762529-e1c5-4358-bc43-c0d81a4c6df6" (UID: "08762529-e1c5-4358-bc43-c0d81a4c6df6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.187942 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08762529-e1c5-4358-bc43-c0d81a4c6df6" (UID: "08762529-e1c5-4358-bc43-c0d81a4c6df6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.200734 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "08762529-e1c5-4358-bc43-c0d81a4c6df6" (UID: "08762529-e1c5-4358-bc43-c0d81a4c6df6"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.242402 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.242460 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.242469 4912 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.242480 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcxrz\" (UniqueName: \"kubernetes.io/projected/08762529-e1c5-4358-bc43-c0d81a4c6df6-kube-api-access-tcxrz\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.242491 4912 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/08762529-e1c5-4358-bc43-c0d81a4c6df6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.242500 4912 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/08762529-e1c5-4358-bc43-c0d81a4c6df6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.315489 4912 scope.go:117] "RemoveContainer" containerID="c77922fad380b1690cfe80e8b1b9cb7b6921b63e8a0da1cdc0b8624aac711282" Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.469678 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d9dc8d44b-7tljg"] Dec 03 01:50:33 crc kubenswrapper[4912]: I1203 01:50:33.479797 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d9dc8d44b-7tljg"] Dec 03 01:50:34 crc kubenswrapper[4912]: I1203 01:50:34.587522 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:50:34 crc kubenswrapper[4912]: E1203 01:50:34.588379 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:50:34 crc kubenswrapper[4912]: I1203 01:50:34.603267 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" path="/var/lib/kubelet/pods/08762529-e1c5-4358-bc43-c0d81a4c6df6/volumes" Dec 03 01:50:35 crc kubenswrapper[4912]: I1203 01:50:35.646496 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 03 01:50:47 crc kubenswrapper[4912]: I1203 01:50:47.572121 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:50:47 crc kubenswrapper[4912]: E1203 01:50:47.573017 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:51:02 crc kubenswrapper[4912]: I1203 01:51:02.573315 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:51:02 crc kubenswrapper[4912]: E1203 01:51:02.574859 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:51:16 crc kubenswrapper[4912]: I1203 01:51:16.572164 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:51:16 crc kubenswrapper[4912]: E1203 01:51:16.573346 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:51:29 crc kubenswrapper[4912]: I1203 01:51:29.571908 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:51:29 crc kubenswrapper[4912]: E1203 01:51:29.572569 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:51:40 crc kubenswrapper[4912]: E1203 01:51:40.384709 4912 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.20:42004->38.102.83.20:41775: write tcp 38.102.83.20:42004->38.102.83.20:41775: write: connection reset by peer Dec 03 01:51:40 crc kubenswrapper[4912]: I1203 01:51:40.572677 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:51:40 crc kubenswrapper[4912]: E1203 01:51:40.573351 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:51:45 crc kubenswrapper[4912]: E1203 01:51:45.785124 4912 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.20:53552->38.102.83.20:41775: write tcp 38.102.83.20:53552->38.102.83.20:41775: write: connection reset by peer Dec 03 01:51:52 crc kubenswrapper[4912]: I1203 01:51:52.575523 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:51:52 crc kubenswrapper[4912]: E1203 01:51:52.576579 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:52:05 crc kubenswrapper[4912]: I1203 01:52:05.572381 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:52:05 crc kubenswrapper[4912]: E1203 01:52:05.573511 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:52:18 crc kubenswrapper[4912]: I1203 01:52:18.572794 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:52:19 crc kubenswrapper[4912]: I1203 01:52:19.571382 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"02bf58e0d4103947f4c9fab9ffca3a905d735d2d6b07f0a24d4cc83c71e28c44"} Dec 03 01:54:18 crc kubenswrapper[4912]: I1203 01:54:18.077904 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:54:18 crc kubenswrapper[4912]: I1203 01:54:18.078577 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:54:21 crc kubenswrapper[4912]: I1203 01:54:21.941407 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2mw4t"] Dec 03 01:54:21 crc kubenswrapper[4912]: E1203 01:54:21.942687 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon" Dec 03 01:54:21 crc kubenswrapper[4912]: I1203 01:54:21.942705 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon" Dec 03 01:54:21 crc kubenswrapper[4912]: E1203 01:54:21.942747 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon-log" Dec 03 01:54:21 crc kubenswrapper[4912]: I1203 01:54:21.942757 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon-log" Dec 03 01:54:21 crc kubenswrapper[4912]: I1203 01:54:21.943062 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon" Dec 03 01:54:21 crc kubenswrapper[4912]: I1203 01:54:21.943154 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="08762529-e1c5-4358-bc43-c0d81a4c6df6" containerName="horizon-log" Dec 03 01:54:21 crc kubenswrapper[4912]: I1203 01:54:21.945495 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:21 crc kubenswrapper[4912]: I1203 01:54:21.978723 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mw4t"] Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.132783 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-utilities\") pod \"redhat-marketplace-2mw4t\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.133836 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-catalog-content\") pod \"redhat-marketplace-2mw4t\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.134037 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnwrt\" (UniqueName: \"kubernetes.io/projected/1877f7f7-72d0-4aff-be86-4835d9d084fb-kube-api-access-jnwrt\") pod \"redhat-marketplace-2mw4t\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.236780 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-utilities\") pod \"redhat-marketplace-2mw4t\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.236923 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-catalog-content\") pod \"redhat-marketplace-2mw4t\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.236958 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnwrt\" (UniqueName: \"kubernetes.io/projected/1877f7f7-72d0-4aff-be86-4835d9d084fb-kube-api-access-jnwrt\") pod \"redhat-marketplace-2mw4t\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.237577 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-utilities\") pod \"redhat-marketplace-2mw4t\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.237855 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-catalog-content\") pod \"redhat-marketplace-2mw4t\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.261863 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnwrt\" (UniqueName: \"kubernetes.io/projected/1877f7f7-72d0-4aff-be86-4835d9d084fb-kube-api-access-jnwrt\") pod \"redhat-marketplace-2mw4t\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.283368 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:22 crc kubenswrapper[4912]: I1203 01:54:22.853324 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mw4t"] Dec 03 01:54:23 crc kubenswrapper[4912]: I1203 01:54:23.254170 4912 generic.go:334] "Generic (PLEG): container finished" podID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerID="de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401" exitCode=0 Dec 03 01:54:23 crc kubenswrapper[4912]: I1203 01:54:23.254225 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mw4t" event={"ID":"1877f7f7-72d0-4aff-be86-4835d9d084fb","Type":"ContainerDied","Data":"de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401"} Dec 03 01:54:23 crc kubenswrapper[4912]: I1203 01:54:23.254258 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mw4t" event={"ID":"1877f7f7-72d0-4aff-be86-4835d9d084fb","Type":"ContainerStarted","Data":"f7360e4bb7d6b8d6ae5ab4b54d2885fd9c0678d44dd36f66337ee66d6c50c7d1"} Dec 03 01:54:24 crc kubenswrapper[4912]: I1203 01:54:24.273043 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mw4t" event={"ID":"1877f7f7-72d0-4aff-be86-4835d9d084fb","Type":"ContainerStarted","Data":"21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4"} Dec 03 01:54:25 crc kubenswrapper[4912]: I1203 01:54:25.288479 4912 generic.go:334] "Generic (PLEG): container finished" podID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerID="21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4" exitCode=0 Dec 03 01:54:25 crc kubenswrapper[4912]: I1203 01:54:25.288548 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mw4t" event={"ID":"1877f7f7-72d0-4aff-be86-4835d9d084fb","Type":"ContainerDied","Data":"21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4"} Dec 03 01:54:26 crc kubenswrapper[4912]: I1203 01:54:26.302314 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mw4t" event={"ID":"1877f7f7-72d0-4aff-be86-4835d9d084fb","Type":"ContainerStarted","Data":"4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50"} Dec 03 01:54:26 crc kubenswrapper[4912]: I1203 01:54:26.342933 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2mw4t" podStartSLOduration=2.925903237 podStartE2EDuration="5.34290936s" podCreationTimestamp="2025-12-03 01:54:21 +0000 UTC" firstStartedPulling="2025-12-03 01:54:23.256364049 +0000 UTC m=+5448.898384609" lastFinishedPulling="2025-12-03 01:54:25.673370132 +0000 UTC m=+5451.315390732" observedRunningTime="2025-12-03 01:54:26.331876193 +0000 UTC m=+5451.973896763" watchObservedRunningTime="2025-12-03 01:54:26.34290936 +0000 UTC m=+5451.984929930" Dec 03 01:54:26 crc kubenswrapper[4912]: I1203 01:54:26.726923 4912 scope.go:117] "RemoveContainer" containerID="cd8e0347ad84a6fc18c9b73f8819b9dc6739c8f29be8cdc0d3f724fb37e6143f" Dec 03 01:54:26 crc kubenswrapper[4912]: I1203 01:54:26.750422 4912 scope.go:117] "RemoveContainer" containerID="046a5a967b2d34f4bdd2f85f053a072a2ef0de3c04c6cdcaf59a6e5f34610bc9" Dec 03 01:54:26 crc kubenswrapper[4912]: I1203 01:54:26.818028 4912 scope.go:117] "RemoveContainer" containerID="7afd0a9e398bb78fed3b7bdda38f9f27420d7c65264353e7634bedd6e408ee14" Dec 03 01:54:32 crc kubenswrapper[4912]: I1203 01:54:32.283974 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:32 crc kubenswrapper[4912]: I1203 01:54:32.284246 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:32 crc kubenswrapper[4912]: I1203 01:54:32.346656 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:32 crc kubenswrapper[4912]: I1203 01:54:32.432630 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:32 crc kubenswrapper[4912]: I1203 01:54:32.610142 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mw4t"] Dec 03 01:54:34 crc kubenswrapper[4912]: I1203 01:54:34.393804 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2mw4t" podUID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerName="registry-server" containerID="cri-o://4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50" gracePeriod=2 Dec 03 01:54:34 crc kubenswrapper[4912]: I1203 01:54:34.989771 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.060906 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnwrt\" (UniqueName: \"kubernetes.io/projected/1877f7f7-72d0-4aff-be86-4835d9d084fb-kube-api-access-jnwrt\") pod \"1877f7f7-72d0-4aff-be86-4835d9d084fb\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.061043 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-catalog-content\") pod \"1877f7f7-72d0-4aff-be86-4835d9d084fb\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.061151 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-utilities\") pod \"1877f7f7-72d0-4aff-be86-4835d9d084fb\" (UID: \"1877f7f7-72d0-4aff-be86-4835d9d084fb\") " Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.061836 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-utilities" (OuterVolumeSpecName: "utilities") pod "1877f7f7-72d0-4aff-be86-4835d9d084fb" (UID: "1877f7f7-72d0-4aff-be86-4835d9d084fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.068394 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1877f7f7-72d0-4aff-be86-4835d9d084fb-kube-api-access-jnwrt" (OuterVolumeSpecName: "kube-api-access-jnwrt") pod "1877f7f7-72d0-4aff-be86-4835d9d084fb" (UID: "1877f7f7-72d0-4aff-be86-4835d9d084fb"). InnerVolumeSpecName "kube-api-access-jnwrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.079339 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1877f7f7-72d0-4aff-be86-4835d9d084fb" (UID: "1877f7f7-72d0-4aff-be86-4835d9d084fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.163460 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnwrt\" (UniqueName: \"kubernetes.io/projected/1877f7f7-72d0-4aff-be86-4835d9d084fb-kube-api-access-jnwrt\") on node \"crc\" DevicePath \"\"" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.163491 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.163502 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1877f7f7-72d0-4aff-be86-4835d9d084fb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.406350 4912 generic.go:334] "Generic (PLEG): container finished" podID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerID="4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50" exitCode=0 Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.406487 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mw4t" event={"ID":"1877f7f7-72d0-4aff-be86-4835d9d084fb","Type":"ContainerDied","Data":"4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50"} Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.406592 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2mw4t" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.406665 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2mw4t" event={"ID":"1877f7f7-72d0-4aff-be86-4835d9d084fb","Type":"ContainerDied","Data":"f7360e4bb7d6b8d6ae5ab4b54d2885fd9c0678d44dd36f66337ee66d6c50c7d1"} Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.406764 4912 scope.go:117] "RemoveContainer" containerID="4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.444694 4912 scope.go:117] "RemoveContainer" containerID="21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.460567 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mw4t"] Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.474244 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2mw4t"] Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.486685 4912 scope.go:117] "RemoveContainer" containerID="de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.526538 4912 scope.go:117] "RemoveContainer" containerID="4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50" Dec 03 01:54:35 crc kubenswrapper[4912]: E1203 01:54:35.527073 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50\": container with ID starting with 4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50 not found: ID does not exist" containerID="4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.527108 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50"} err="failed to get container status \"4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50\": rpc error: code = NotFound desc = could not find container \"4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50\": container with ID starting with 4f80405c4e5df9db7d814ce176b970a488ff24cc4e58657c79db52da726fdd50 not found: ID does not exist" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.527134 4912 scope.go:117] "RemoveContainer" containerID="21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4" Dec 03 01:54:35 crc kubenswrapper[4912]: E1203 01:54:35.527705 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4\": container with ID starting with 21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4 not found: ID does not exist" containerID="21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.527738 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4"} err="failed to get container status \"21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4\": rpc error: code = NotFound desc = could not find container \"21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4\": container with ID starting with 21450b95cfb0079b91d9bdc9ff0879536be58c3b15af45f2464530f5779f6df4 not found: ID does not exist" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.527756 4912 scope.go:117] "RemoveContainer" containerID="de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401" Dec 03 01:54:35 crc kubenswrapper[4912]: E1203 01:54:35.528013 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401\": container with ID starting with de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401 not found: ID does not exist" containerID="de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401" Dec 03 01:54:35 crc kubenswrapper[4912]: I1203 01:54:35.528050 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401"} err="failed to get container status \"de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401\": rpc error: code = NotFound desc = could not find container \"de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401\": container with ID starting with de1defb98521afa8842875c76c8f592e6f10dfa5950383defb89321452fc0401 not found: ID does not exist" Dec 03 01:54:36 crc kubenswrapper[4912]: I1203 01:54:36.583824 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1877f7f7-72d0-4aff-be86-4835d9d084fb" path="/var/lib/kubelet/pods/1877f7f7-72d0-4aff-be86-4835d9d084fb/volumes" Dec 03 01:54:48 crc kubenswrapper[4912]: I1203 01:54:48.077887 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:54:48 crc kubenswrapper[4912]: I1203 01:54:48.078607 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.709222 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4bvqq"] Dec 03 01:55:15 crc kubenswrapper[4912]: E1203 01:55:15.711178 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerName="extract-utilities" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.711209 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerName="extract-utilities" Dec 03 01:55:15 crc kubenswrapper[4912]: E1203 01:55:15.711234 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerName="extract-content" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.711242 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerName="extract-content" Dec 03 01:55:15 crc kubenswrapper[4912]: E1203 01:55:15.711276 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerName="registry-server" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.711286 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerName="registry-server" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.712060 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="1877f7f7-72d0-4aff-be86-4835d9d084fb" containerName="registry-server" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.717511 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.749975 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bvqq"] Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.823000 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n7kv\" (UniqueName: \"kubernetes.io/projected/0ddcd602-8c96-4fae-9723-47aa0488196c-kube-api-access-9n7kv\") pod \"redhat-operators-4bvqq\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.823171 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-utilities\") pod \"redhat-operators-4bvqq\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.823195 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-catalog-content\") pod \"redhat-operators-4bvqq\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.925271 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n7kv\" (UniqueName: \"kubernetes.io/projected/0ddcd602-8c96-4fae-9723-47aa0488196c-kube-api-access-9n7kv\") pod \"redhat-operators-4bvqq\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.925507 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-utilities\") pod \"redhat-operators-4bvqq\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.925534 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-catalog-content\") pod \"redhat-operators-4bvqq\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.926019 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-catalog-content\") pod \"redhat-operators-4bvqq\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.926055 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-utilities\") pod \"redhat-operators-4bvqq\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:15 crc kubenswrapper[4912]: I1203 01:55:15.945380 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n7kv\" (UniqueName: \"kubernetes.io/projected/0ddcd602-8c96-4fae-9723-47aa0488196c-kube-api-access-9n7kv\") pod \"redhat-operators-4bvqq\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:16 crc kubenswrapper[4912]: I1203 01:55:16.049911 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:16 crc kubenswrapper[4912]: I1203 01:55:16.588747 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bvqq"] Dec 03 01:55:16 crc kubenswrapper[4912]: I1203 01:55:16.923986 4912 generic.go:334] "Generic (PLEG): container finished" podID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerID="e9817af3e535a132b43570b83f07f9b5395ef9bdfc8ea86b84f0106d5106dc43" exitCode=0 Dec 03 01:55:16 crc kubenswrapper[4912]: I1203 01:55:16.924038 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bvqq" event={"ID":"0ddcd602-8c96-4fae-9723-47aa0488196c","Type":"ContainerDied","Data":"e9817af3e535a132b43570b83f07f9b5395ef9bdfc8ea86b84f0106d5106dc43"} Dec 03 01:55:16 crc kubenswrapper[4912]: I1203 01:55:16.924270 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bvqq" event={"ID":"0ddcd602-8c96-4fae-9723-47aa0488196c","Type":"ContainerStarted","Data":"2e570714a448e52261d85bbd814764c1f68610346192edb750498101fa029d23"} Dec 03 01:55:16 crc kubenswrapper[4912]: I1203 01:55:16.925985 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 01:55:17 crc kubenswrapper[4912]: I1203 01:55:17.936212 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bvqq" event={"ID":"0ddcd602-8c96-4fae-9723-47aa0488196c","Type":"ContainerStarted","Data":"902701276cd5f15db3440e818f922ec943bed7a260b19e6e7f98d76fc6053e35"} Dec 03 01:55:18 crc kubenswrapper[4912]: I1203 01:55:18.077811 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:55:18 crc kubenswrapper[4912]: I1203 01:55:18.077884 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:55:18 crc kubenswrapper[4912]: I1203 01:55:18.077967 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:55:18 crc kubenswrapper[4912]: I1203 01:55:18.079074 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"02bf58e0d4103947f4c9fab9ffca3a905d735d2d6b07f0a24d4cc83c71e28c44"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:55:18 crc kubenswrapper[4912]: I1203 01:55:18.079174 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://02bf58e0d4103947f4c9fab9ffca3a905d735d2d6b07f0a24d4cc83c71e28c44" gracePeriod=600 Dec 03 01:55:18 crc kubenswrapper[4912]: I1203 01:55:18.952672 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="02bf58e0d4103947f4c9fab9ffca3a905d735d2d6b07f0a24d4cc83c71e28c44" exitCode=0 Dec 03 01:55:18 crc kubenswrapper[4912]: I1203 01:55:18.952742 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"02bf58e0d4103947f4c9fab9ffca3a905d735d2d6b07f0a24d4cc83c71e28c44"} Dec 03 01:55:18 crc kubenswrapper[4912]: I1203 01:55:18.953176 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879"} Dec 03 01:55:18 crc kubenswrapper[4912]: I1203 01:55:18.953211 4912 scope.go:117] "RemoveContainer" containerID="9cfa79bcd07d46678f63a79966353b70270540c8ab0a61b9e1bc8a86afbed268" Dec 03 01:55:20 crc kubenswrapper[4912]: I1203 01:55:20.984424 4912 generic.go:334] "Generic (PLEG): container finished" podID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerID="902701276cd5f15db3440e818f922ec943bed7a260b19e6e7f98d76fc6053e35" exitCode=0 Dec 03 01:55:20 crc kubenswrapper[4912]: I1203 01:55:20.984529 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bvqq" event={"ID":"0ddcd602-8c96-4fae-9723-47aa0488196c","Type":"ContainerDied","Data":"902701276cd5f15db3440e818f922ec943bed7a260b19e6e7f98d76fc6053e35"} Dec 03 01:55:22 crc kubenswrapper[4912]: E1203 01:55:22.485842 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/b8/b856e4d37af238240aaa3504ebf72881a05d3e5875365377d4fbd3a313fe7d06?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251203%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251203T015521Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=1132c338798ee4c4ef250ef4fc41e714c3e7735f4f4e7302986ddc28f07ed692®ion=us-east-1&namespace=openshift-release-dev&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=ocp-v4.0-art-dev&akamai_signature=exp=1764727821~hmac=8f8f77eb2d72a7fb8429fa40df8f59496bc5c4e13c56d76f65ed12d3dbd9f300\": remote error: tls: internal error" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad" Dec 03 01:55:22 crc kubenswrapper[4912]: E1203 01:55:22.488825 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad,Command:[/bin/opm],Args:[serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOMEMLIMIT,Value:30MiB,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{31457280 0} {} 30Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9n7kv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4bvqq_openshift-marketplace(0ddcd602-8c96-4fae-9723-47aa0488196c): ErrImagePull: parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/b8/b856e4d37af238240aaa3504ebf72881a05d3e5875365377d4fbd3a313fe7d06?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251203%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251203T015521Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=1132c338798ee4c4ef250ef4fc41e714c3e7735f4f4e7302986ddc28f07ed692®ion=us-east-1&namespace=openshift-release-dev&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=ocp-v4.0-art-dev&akamai_signature=exp=1764727821~hmac=8f8f77eb2d72a7fb8429fa40df8f59496bc5c4e13c56d76f65ed12d3dbd9f300\": remote error: tls: internal error" logger="UnhandledError" Dec 03 01:55:22 crc kubenswrapper[4912]: E1203 01:55:22.490093 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"parsing image configuration: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/b8/b856e4d37af238240aaa3504ebf72881a05d3e5875365377d4fbd3a313fe7d06?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251203%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251203T015521Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=1132c338798ee4c4ef250ef4fc41e714c3e7735f4f4e7302986ddc28f07ed692®ion=us-east-1&namespace=openshift-release-dev&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=ocp-v4.0-art-dev&akamai_signature=exp=1764727821~hmac=8f8f77eb2d72a7fb8429fa40df8f59496bc5c4e13c56d76f65ed12d3dbd9f300\\\": remote error: tls: internal error\"" pod="openshift-marketplace/redhat-operators-4bvqq" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" Dec 03 01:55:36 crc kubenswrapper[4912]: I1203 01:55:36.210600 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bvqq" event={"ID":"0ddcd602-8c96-4fae-9723-47aa0488196c","Type":"ContainerStarted","Data":"df39ea93e56af414f54b36cc4f6ae12545df67e5b8bc2508707e8d3ced75fa57"} Dec 03 01:55:36 crc kubenswrapper[4912]: I1203 01:55:36.248643 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4bvqq" podStartSLOduration=3.101806301 podStartE2EDuration="21.248621648s" podCreationTimestamp="2025-12-03 01:55:15 +0000 UTC" firstStartedPulling="2025-12-03 01:55:16.925762862 +0000 UTC m=+5502.567783422" lastFinishedPulling="2025-12-03 01:55:35.072578169 +0000 UTC m=+5520.714598769" observedRunningTime="2025-12-03 01:55:36.23753889 +0000 UTC m=+5521.879559500" watchObservedRunningTime="2025-12-03 01:55:36.248621648 +0000 UTC m=+5521.890642208" Dec 03 01:55:37 crc kubenswrapper[4912]: E1203 01:55:37.818418 4912 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.20:38776->38.102.83.20:41775: write tcp 38.102.83.20:38776->38.102.83.20:41775: write: connection reset by peer Dec 03 01:55:46 crc kubenswrapper[4912]: I1203 01:55:46.050044 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:46 crc kubenswrapper[4912]: I1203 01:55:46.050702 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:46 crc kubenswrapper[4912]: I1203 01:55:46.113999 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:46 crc kubenswrapper[4912]: I1203 01:55:46.413997 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:49 crc kubenswrapper[4912]: I1203 01:55:49.248853 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bvqq"] Dec 03 01:55:49 crc kubenswrapper[4912]: I1203 01:55:49.249918 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4bvqq" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerName="registry-server" containerID="cri-o://df39ea93e56af414f54b36cc4f6ae12545df67e5b8bc2508707e8d3ced75fa57" gracePeriod=2 Dec 03 01:55:50 crc kubenswrapper[4912]: I1203 01:55:50.375832 4912 generic.go:334] "Generic (PLEG): container finished" podID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerID="df39ea93e56af414f54b36cc4f6ae12545df67e5b8bc2508707e8d3ced75fa57" exitCode=0 Dec 03 01:55:50 crc kubenswrapper[4912]: I1203 01:55:50.375881 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bvqq" event={"ID":"0ddcd602-8c96-4fae-9723-47aa0488196c","Type":"ContainerDied","Data":"df39ea93e56af414f54b36cc4f6ae12545df67e5b8bc2508707e8d3ced75fa57"} Dec 03 01:55:50 crc kubenswrapper[4912]: I1203 01:55:50.843684 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:50 crc kubenswrapper[4912]: I1203 01:55:50.958904 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-utilities\") pod \"0ddcd602-8c96-4fae-9723-47aa0488196c\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " Dec 03 01:55:50 crc kubenswrapper[4912]: I1203 01:55:50.959061 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-catalog-content\") pod \"0ddcd602-8c96-4fae-9723-47aa0488196c\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " Dec 03 01:55:50 crc kubenswrapper[4912]: I1203 01:55:50.959216 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n7kv\" (UniqueName: \"kubernetes.io/projected/0ddcd602-8c96-4fae-9723-47aa0488196c-kube-api-access-9n7kv\") pod \"0ddcd602-8c96-4fae-9723-47aa0488196c\" (UID: \"0ddcd602-8c96-4fae-9723-47aa0488196c\") " Dec 03 01:55:50 crc kubenswrapper[4912]: I1203 01:55:50.959630 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-utilities" (OuterVolumeSpecName: "utilities") pod "0ddcd602-8c96-4fae-9723-47aa0488196c" (UID: "0ddcd602-8c96-4fae-9723-47aa0488196c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:55:50 crc kubenswrapper[4912]: I1203 01:55:50.959795 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:55:50 crc kubenswrapper[4912]: I1203 01:55:50.964779 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ddcd602-8c96-4fae-9723-47aa0488196c-kube-api-access-9n7kv" (OuterVolumeSpecName: "kube-api-access-9n7kv") pod "0ddcd602-8c96-4fae-9723-47aa0488196c" (UID: "0ddcd602-8c96-4fae-9723-47aa0488196c"). InnerVolumeSpecName "kube-api-access-9n7kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.062647 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n7kv\" (UniqueName: \"kubernetes.io/projected/0ddcd602-8c96-4fae-9723-47aa0488196c-kube-api-access-9n7kv\") on node \"crc\" DevicePath \"\"" Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.073045 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ddcd602-8c96-4fae-9723-47aa0488196c" (UID: "0ddcd602-8c96-4fae-9723-47aa0488196c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.164654 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ddcd602-8c96-4fae-9723-47aa0488196c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.393943 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bvqq" event={"ID":"0ddcd602-8c96-4fae-9723-47aa0488196c","Type":"ContainerDied","Data":"2e570714a448e52261d85bbd814764c1f68610346192edb750498101fa029d23"} Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.394016 4912 scope.go:117] "RemoveContainer" containerID="df39ea93e56af414f54b36cc4f6ae12545df67e5b8bc2508707e8d3ced75fa57" Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.394265 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bvqq" Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.431307 4912 scope.go:117] "RemoveContainer" containerID="902701276cd5f15db3440e818f922ec943bed7a260b19e6e7f98d76fc6053e35" Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.446740 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bvqq"] Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.460410 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4bvqq"] Dec 03 01:55:51 crc kubenswrapper[4912]: I1203 01:55:51.724400 4912 scope.go:117] "RemoveContainer" containerID="e9817af3e535a132b43570b83f07f9b5395ef9bdfc8ea86b84f0106d5106dc43" Dec 03 01:55:52 crc kubenswrapper[4912]: I1203 01:55:52.588151 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" path="/var/lib/kubelet/pods/0ddcd602-8c96-4fae-9723-47aa0488196c/volumes" Dec 03 01:57:18 crc kubenswrapper[4912]: I1203 01:57:18.077917 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:57:18 crc kubenswrapper[4912]: I1203 01:57:18.078743 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:57:48 crc kubenswrapper[4912]: I1203 01:57:48.077605 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:57:48 crc kubenswrapper[4912]: I1203 01:57:48.078758 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.713196 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2dhqd"] Dec 03 01:58:14 crc kubenswrapper[4912]: E1203 01:58:14.714191 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerName="extract-utilities" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.714205 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerName="extract-utilities" Dec 03 01:58:14 crc kubenswrapper[4912]: E1203 01:58:14.714235 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerName="registry-server" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.714242 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerName="registry-server" Dec 03 01:58:14 crc kubenswrapper[4912]: E1203 01:58:14.714266 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerName="extract-content" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.714272 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerName="extract-content" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.714589 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ddcd602-8c96-4fae-9723-47aa0488196c" containerName="registry-server" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.716154 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.755025 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88h2m\" (UniqueName: \"kubernetes.io/projected/1620a740-28ff-47c0-98bd-f80540945d44-kube-api-access-88h2m\") pod \"certified-operators-2dhqd\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.755159 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-catalog-content\") pod \"certified-operators-2dhqd\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.755288 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-utilities\") pod \"certified-operators-2dhqd\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.819157 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dhqd"] Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.864512 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88h2m\" (UniqueName: \"kubernetes.io/projected/1620a740-28ff-47c0-98bd-f80540945d44-kube-api-access-88h2m\") pod \"certified-operators-2dhqd\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.864683 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-catalog-content\") pod \"certified-operators-2dhqd\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.865144 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-catalog-content\") pod \"certified-operators-2dhqd\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.865284 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-utilities\") pod \"certified-operators-2dhqd\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.865713 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-utilities\") pod \"certified-operators-2dhqd\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:14 crc kubenswrapper[4912]: I1203 01:58:14.901016 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88h2m\" (UniqueName: \"kubernetes.io/projected/1620a740-28ff-47c0-98bd-f80540945d44-kube-api-access-88h2m\") pod \"certified-operators-2dhqd\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:15 crc kubenswrapper[4912]: I1203 01:58:15.040094 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:15 crc kubenswrapper[4912]: I1203 01:58:15.495898 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dhqd"] Dec 03 01:58:16 crc kubenswrapper[4912]: I1203 01:58:16.474302 4912 generic.go:334] "Generic (PLEG): container finished" podID="1620a740-28ff-47c0-98bd-f80540945d44" containerID="e78e04d8ce169800429b4959bb1601181d3eee2073434e851f534bfbddc196da" exitCode=0 Dec 03 01:58:16 crc kubenswrapper[4912]: I1203 01:58:16.474448 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dhqd" event={"ID":"1620a740-28ff-47c0-98bd-f80540945d44","Type":"ContainerDied","Data":"e78e04d8ce169800429b4959bb1601181d3eee2073434e851f534bfbddc196da"} Dec 03 01:58:16 crc kubenswrapper[4912]: I1203 01:58:16.474847 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dhqd" event={"ID":"1620a740-28ff-47c0-98bd-f80540945d44","Type":"ContainerStarted","Data":"1783e1c6c00bb5fc2def85a167678ba0b0002d4386b90cfe8947cb6e36b7a667"} Dec 03 01:58:17 crc kubenswrapper[4912]: I1203 01:58:17.489930 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dhqd" event={"ID":"1620a740-28ff-47c0-98bd-f80540945d44","Type":"ContainerStarted","Data":"6fdac02cfbc48bc905dececf383e17a6c0e0aa849b9a43741de670f0b50bba2b"} Dec 03 01:58:18 crc kubenswrapper[4912]: I1203 01:58:18.077500 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 01:58:18 crc kubenswrapper[4912]: I1203 01:58:18.077576 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 01:58:18 crc kubenswrapper[4912]: I1203 01:58:18.077633 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 01:58:18 crc kubenswrapper[4912]: I1203 01:58:18.078423 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 01:58:18 crc kubenswrapper[4912]: I1203 01:58:18.078549 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" gracePeriod=600 Dec 03 01:58:18 crc kubenswrapper[4912]: E1203 01:58:18.917039 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:58:19 crc kubenswrapper[4912]: I1203 01:58:19.517648 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" exitCode=0 Dec 03 01:58:19 crc kubenswrapper[4912]: I1203 01:58:19.518088 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879"} Dec 03 01:58:19 crc kubenswrapper[4912]: I1203 01:58:19.518144 4912 scope.go:117] "RemoveContainer" containerID="02bf58e0d4103947f4c9fab9ffca3a905d735d2d6b07f0a24d4cc83c71e28c44" Dec 03 01:58:19 crc kubenswrapper[4912]: I1203 01:58:19.519220 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 01:58:19 crc kubenswrapper[4912]: E1203 01:58:19.519847 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:58:19 crc kubenswrapper[4912]: I1203 01:58:19.528021 4912 generic.go:334] "Generic (PLEG): container finished" podID="1620a740-28ff-47c0-98bd-f80540945d44" containerID="6fdac02cfbc48bc905dececf383e17a6c0e0aa849b9a43741de670f0b50bba2b" exitCode=0 Dec 03 01:58:19 crc kubenswrapper[4912]: I1203 01:58:19.528087 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dhqd" event={"ID":"1620a740-28ff-47c0-98bd-f80540945d44","Type":"ContainerDied","Data":"6fdac02cfbc48bc905dececf383e17a6c0e0aa849b9a43741de670f0b50bba2b"} Dec 03 01:58:20 crc kubenswrapper[4912]: I1203 01:58:20.543048 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dhqd" event={"ID":"1620a740-28ff-47c0-98bd-f80540945d44","Type":"ContainerStarted","Data":"8f794d9af644b97d8050fb79aa33c495446fabe71c8651368b8527164fdb3e8b"} Dec 03 01:58:20 crc kubenswrapper[4912]: I1203 01:58:20.594827 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2dhqd" podStartSLOduration=3.081825638 podStartE2EDuration="6.594798666s" podCreationTimestamp="2025-12-03 01:58:14 +0000 UTC" firstStartedPulling="2025-12-03 01:58:16.47771111 +0000 UTC m=+5682.119731690" lastFinishedPulling="2025-12-03 01:58:19.990684158 +0000 UTC m=+5685.632704718" observedRunningTime="2025-12-03 01:58:20.572357772 +0000 UTC m=+5686.214378322" watchObservedRunningTime="2025-12-03 01:58:20.594798666 +0000 UTC m=+5686.236819236" Dec 03 01:58:25 crc kubenswrapper[4912]: I1203 01:58:25.040453 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:25 crc kubenswrapper[4912]: I1203 01:58:25.040797 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:25 crc kubenswrapper[4912]: I1203 01:58:25.107580 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:25 crc kubenswrapper[4912]: I1203 01:58:25.752766 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:25 crc kubenswrapper[4912]: I1203 01:58:25.860157 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dhqd"] Dec 03 01:58:27 crc kubenswrapper[4912]: I1203 01:58:27.640569 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2dhqd" podUID="1620a740-28ff-47c0-98bd-f80540945d44" containerName="registry-server" containerID="cri-o://8f794d9af644b97d8050fb79aa33c495446fabe71c8651368b8527164fdb3e8b" gracePeriod=2 Dec 03 01:58:28 crc kubenswrapper[4912]: I1203 01:58:28.655655 4912 generic.go:334] "Generic (PLEG): container finished" podID="1620a740-28ff-47c0-98bd-f80540945d44" containerID="8f794d9af644b97d8050fb79aa33c495446fabe71c8651368b8527164fdb3e8b" exitCode=0 Dec 03 01:58:28 crc kubenswrapper[4912]: I1203 01:58:28.655992 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dhqd" event={"ID":"1620a740-28ff-47c0-98bd-f80540945d44","Type":"ContainerDied","Data":"8f794d9af644b97d8050fb79aa33c495446fabe71c8651368b8527164fdb3e8b"} Dec 03 01:58:28 crc kubenswrapper[4912]: I1203 01:58:28.973781 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.023590 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88h2m\" (UniqueName: \"kubernetes.io/projected/1620a740-28ff-47c0-98bd-f80540945d44-kube-api-access-88h2m\") pod \"1620a740-28ff-47c0-98bd-f80540945d44\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.023699 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-catalog-content\") pod \"1620a740-28ff-47c0-98bd-f80540945d44\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.023810 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-utilities\") pod \"1620a740-28ff-47c0-98bd-f80540945d44\" (UID: \"1620a740-28ff-47c0-98bd-f80540945d44\") " Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.024709 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-utilities" (OuterVolumeSpecName: "utilities") pod "1620a740-28ff-47c0-98bd-f80540945d44" (UID: "1620a740-28ff-47c0-98bd-f80540945d44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.030787 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1620a740-28ff-47c0-98bd-f80540945d44-kube-api-access-88h2m" (OuterVolumeSpecName: "kube-api-access-88h2m") pod "1620a740-28ff-47c0-98bd-f80540945d44" (UID: "1620a740-28ff-47c0-98bd-f80540945d44"). InnerVolumeSpecName "kube-api-access-88h2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.092124 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1620a740-28ff-47c0-98bd-f80540945d44" (UID: "1620a740-28ff-47c0-98bd-f80540945d44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.126044 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88h2m\" (UniqueName: \"kubernetes.io/projected/1620a740-28ff-47c0-98bd-f80540945d44-kube-api-access-88h2m\") on node \"crc\" DevicePath \"\"" Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.126086 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.126098 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1620a740-28ff-47c0-98bd-f80540945d44-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.678991 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dhqd" event={"ID":"1620a740-28ff-47c0-98bd-f80540945d44","Type":"ContainerDied","Data":"1783e1c6c00bb5fc2def85a167678ba0b0002d4386b90cfe8947cb6e36b7a667"} Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.679288 4912 scope.go:117] "RemoveContainer" containerID="8f794d9af644b97d8050fb79aa33c495446fabe71c8651368b8527164fdb3e8b" Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.679315 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dhqd" Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.715545 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dhqd"] Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.722208 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2dhqd"] Dec 03 01:58:29 crc kubenswrapper[4912]: I1203 01:58:29.737982 4912 scope.go:117] "RemoveContainer" containerID="6fdac02cfbc48bc905dececf383e17a6c0e0aa849b9a43741de670f0b50bba2b" Dec 03 01:58:30 crc kubenswrapper[4912]: I1203 01:58:30.512493 4912 scope.go:117] "RemoveContainer" containerID="e78e04d8ce169800429b4959bb1601181d3eee2073434e851f534bfbddc196da" Dec 03 01:58:30 crc kubenswrapper[4912]: I1203 01:58:30.577624 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 01:58:30 crc kubenswrapper[4912]: E1203 01:58:30.578150 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:58:30 crc kubenswrapper[4912]: I1203 01:58:30.587988 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1620a740-28ff-47c0-98bd-f80540945d44" path="/var/lib/kubelet/pods/1620a740-28ff-47c0-98bd-f80540945d44/volumes" Dec 03 01:58:44 crc kubenswrapper[4912]: I1203 01:58:44.590275 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 01:58:44 crc kubenswrapper[4912]: E1203 01:58:44.591652 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:58:57 crc kubenswrapper[4912]: I1203 01:58:57.572351 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 01:58:57 crc kubenswrapper[4912]: E1203 01:58:57.573586 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:59:07 crc kubenswrapper[4912]: I1203 01:59:07.103504 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-jg4rj"] Dec 03 01:59:07 crc kubenswrapper[4912]: I1203 01:59:07.119144 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-d0c6-account-create-update-r2rnf"] Dec 03 01:59:07 crc kubenswrapper[4912]: I1203 01:59:07.135937 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-jg4rj"] Dec 03 01:59:07 crc kubenswrapper[4912]: I1203 01:59:07.151708 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-d0c6-account-create-update-r2rnf"] Dec 03 01:59:08 crc kubenswrapper[4912]: I1203 01:59:08.627344 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69661ff8-038d-4f96-b806-877ec5aaa6ba" path="/var/lib/kubelet/pods/69661ff8-038d-4f96-b806-877ec5aaa6ba/volumes" Dec 03 01:59:08 crc kubenswrapper[4912]: I1203 01:59:08.629105 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86aa696b-0e24-4cde-9944-4142d0fb81a8" path="/var/lib/kubelet/pods/86aa696b-0e24-4cde-9944-4142d0fb81a8/volumes" Dec 03 01:59:12 crc kubenswrapper[4912]: I1203 01:59:12.572545 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 01:59:12 crc kubenswrapper[4912]: E1203 01:59:12.573638 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:59:26 crc kubenswrapper[4912]: I1203 01:59:26.573860 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 01:59:26 crc kubenswrapper[4912]: E1203 01:59:26.575015 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:59:27 crc kubenswrapper[4912]: I1203 01:59:27.054166 4912 scope.go:117] "RemoveContainer" containerID="783a6eb0f0944208070d0e047330fdb365a5e286bcb47bb82a99a5e0f047230b" Dec 03 01:59:27 crc kubenswrapper[4912]: I1203 01:59:27.965645 4912 scope.go:117] "RemoveContainer" containerID="4afcfb2b941810d05bb5ba3307bfe70475956503b8913190ff8e426fb6791268" Dec 03 01:59:38 crc kubenswrapper[4912]: I1203 01:59:38.038215 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-h7pdq"] Dec 03 01:59:38 crc kubenswrapper[4912]: I1203 01:59:38.046829 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-h7pdq"] Dec 03 01:59:38 crc kubenswrapper[4912]: I1203 01:59:38.597771 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe7da98f-555b-47d1-ae2a-849d69e7540a" path="/var/lib/kubelet/pods/fe7da98f-555b-47d1-ae2a-849d69e7540a/volumes" Dec 03 01:59:41 crc kubenswrapper[4912]: I1203 01:59:41.572726 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 01:59:41 crc kubenswrapper[4912]: E1203 01:59:41.573606 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.294642 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cz2nd"] Dec 03 01:59:53 crc kubenswrapper[4912]: E1203 01:59:53.296022 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1620a740-28ff-47c0-98bd-f80540945d44" containerName="registry-server" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.296043 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1620a740-28ff-47c0-98bd-f80540945d44" containerName="registry-server" Dec 03 01:59:53 crc kubenswrapper[4912]: E1203 01:59:53.296095 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1620a740-28ff-47c0-98bd-f80540945d44" containerName="extract-content" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.296105 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1620a740-28ff-47c0-98bd-f80540945d44" containerName="extract-content" Dec 03 01:59:53 crc kubenswrapper[4912]: E1203 01:59:53.296121 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1620a740-28ff-47c0-98bd-f80540945d44" containerName="extract-utilities" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.296130 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="1620a740-28ff-47c0-98bd-f80540945d44" containerName="extract-utilities" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.296407 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="1620a740-28ff-47c0-98bd-f80540945d44" containerName="registry-server" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.299804 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.323039 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cz2nd"] Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.377660 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkrrx\" (UniqueName: \"kubernetes.io/projected/244bf195-a521-4253-bb1d-52105460dd83-kube-api-access-pkrrx\") pod \"community-operators-cz2nd\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.377745 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-utilities\") pod \"community-operators-cz2nd\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.377835 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-catalog-content\") pod \"community-operators-cz2nd\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.479836 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkrrx\" (UniqueName: \"kubernetes.io/projected/244bf195-a521-4253-bb1d-52105460dd83-kube-api-access-pkrrx\") pod \"community-operators-cz2nd\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.479916 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-utilities\") pod \"community-operators-cz2nd\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.480026 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-catalog-content\") pod \"community-operators-cz2nd\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.480594 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-catalog-content\") pod \"community-operators-cz2nd\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.480657 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-utilities\") pod \"community-operators-cz2nd\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.536566 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkrrx\" (UniqueName: \"kubernetes.io/projected/244bf195-a521-4253-bb1d-52105460dd83-kube-api-access-pkrrx\") pod \"community-operators-cz2nd\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:53 crc kubenswrapper[4912]: I1203 01:59:53.633218 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cz2nd" Dec 03 01:59:54 crc kubenswrapper[4912]: I1203 01:59:54.190274 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cz2nd"] Dec 03 01:59:54 crc kubenswrapper[4912]: I1203 01:59:54.808170 4912 generic.go:334] "Generic (PLEG): container finished" podID="244bf195-a521-4253-bb1d-52105460dd83" containerID="1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288" exitCode=0 Dec 03 01:59:54 crc kubenswrapper[4912]: I1203 01:59:54.808218 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cz2nd" event={"ID":"244bf195-a521-4253-bb1d-52105460dd83","Type":"ContainerDied","Data":"1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288"} Dec 03 01:59:54 crc kubenswrapper[4912]: I1203 01:59:54.808644 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cz2nd" event={"ID":"244bf195-a521-4253-bb1d-52105460dd83","Type":"ContainerStarted","Data":"99808ff95106da96d75e62caa54e70836ad5002c8d22f3363bcc67410a34a6f6"} Dec 03 01:59:55 crc kubenswrapper[4912]: I1203 01:59:55.826994 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cz2nd" event={"ID":"244bf195-a521-4253-bb1d-52105460dd83","Type":"ContainerStarted","Data":"4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b"} Dec 03 01:59:56 crc kubenswrapper[4912]: I1203 01:59:56.572108 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 01:59:56 crc kubenswrapper[4912]: E1203 01:59:56.572838 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 01:59:56 crc kubenswrapper[4912]: I1203 01:59:56.862264 4912 generic.go:334] "Generic (PLEG): container finished" podID="244bf195-a521-4253-bb1d-52105460dd83" containerID="4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b" exitCode=0 Dec 03 01:59:56 crc kubenswrapper[4912]: I1203 01:59:56.862334 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cz2nd" event={"ID":"244bf195-a521-4253-bb1d-52105460dd83","Type":"ContainerDied","Data":"4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b"} Dec 03 01:59:57 crc kubenswrapper[4912]: I1203 01:59:57.878836 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cz2nd" event={"ID":"244bf195-a521-4253-bb1d-52105460dd83","Type":"ContainerStarted","Data":"fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd"} Dec 03 01:59:57 crc kubenswrapper[4912]: I1203 01:59:57.918824 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cz2nd" podStartSLOduration=2.407079912 podStartE2EDuration="4.918795574s" podCreationTimestamp="2025-12-03 01:59:53 +0000 UTC" firstStartedPulling="2025-12-03 01:59:54.811273848 +0000 UTC m=+5780.453294408" lastFinishedPulling="2025-12-03 01:59:57.32298951 +0000 UTC m=+5782.965010070" observedRunningTime="2025-12-03 01:59:57.903564804 +0000 UTC m=+5783.545585424" watchObservedRunningTime="2025-12-03 01:59:57.918795574 +0000 UTC m=+5783.560816164" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.179397 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj"] Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.182157 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.184491 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.184783 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.191058 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj"] Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.263955 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc9g9\" (UniqueName: \"kubernetes.io/projected/63b375e9-d94c-4e34-882c-be534f9c046b-kube-api-access-xc9g9\") pod \"collect-profiles-29412120-p6phj\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.264030 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63b375e9-d94c-4e34-882c-be534f9c046b-secret-volume\") pod \"collect-profiles-29412120-p6phj\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.264172 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63b375e9-d94c-4e34-882c-be534f9c046b-config-volume\") pod \"collect-profiles-29412120-p6phj\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.366419 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63b375e9-d94c-4e34-882c-be534f9c046b-config-volume\") pod \"collect-profiles-29412120-p6phj\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.366698 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc9g9\" (UniqueName: \"kubernetes.io/projected/63b375e9-d94c-4e34-882c-be534f9c046b-kube-api-access-xc9g9\") pod \"collect-profiles-29412120-p6phj\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.366772 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63b375e9-d94c-4e34-882c-be534f9c046b-secret-volume\") pod \"collect-profiles-29412120-p6phj\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.367531 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63b375e9-d94c-4e34-882c-be534f9c046b-config-volume\") pod \"collect-profiles-29412120-p6phj\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.374332 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63b375e9-d94c-4e34-882c-be534f9c046b-secret-volume\") pod \"collect-profiles-29412120-p6phj\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.384296 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc9g9\" (UniqueName: \"kubernetes.io/projected/63b375e9-d94c-4e34-882c-be534f9c046b-kube-api-access-xc9g9\") pod \"collect-profiles-29412120-p6phj\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:00 crc kubenswrapper[4912]: I1203 02:00:00.518726 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:01 crc kubenswrapper[4912]: I1203 02:00:01.060214 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj"] Dec 03 02:00:01 crc kubenswrapper[4912]: I1203 02:00:01.922228 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" event={"ID":"63b375e9-d94c-4e34-882c-be534f9c046b","Type":"ContainerStarted","Data":"ac1f98654a45487066dc62ce78408749a87794b67926a0e35fe1f98b0d8b684f"} Dec 03 02:00:02 crc kubenswrapper[4912]: I1203 02:00:02.937734 4912 generic.go:334] "Generic (PLEG): container finished" podID="63b375e9-d94c-4e34-882c-be534f9c046b" containerID="d0cb19268c64db4fbee75405ec47954865f2fc9600efd9186bf7a98283a0172a" exitCode=0 Dec 03 02:00:02 crc kubenswrapper[4912]: I1203 02:00:02.937838 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" event={"ID":"63b375e9-d94c-4e34-882c-be534f9c046b","Type":"ContainerDied","Data":"d0cb19268c64db4fbee75405ec47954865f2fc9600efd9186bf7a98283a0172a"} Dec 03 02:00:03 crc kubenswrapper[4912]: I1203 02:00:03.634175 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cz2nd" Dec 03 02:00:03 crc kubenswrapper[4912]: I1203 02:00:03.634560 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cz2nd" Dec 03 02:00:03 crc kubenswrapper[4912]: I1203 02:00:03.723615 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cz2nd" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.001868 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cz2nd" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.068724 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cz2nd"] Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.449821 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.565242 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63b375e9-d94c-4e34-882c-be534f9c046b-secret-volume\") pod \"63b375e9-d94c-4e34-882c-be534f9c046b\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.565629 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63b375e9-d94c-4e34-882c-be534f9c046b-config-volume\") pod \"63b375e9-d94c-4e34-882c-be534f9c046b\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.565756 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc9g9\" (UniqueName: \"kubernetes.io/projected/63b375e9-d94c-4e34-882c-be534f9c046b-kube-api-access-xc9g9\") pod \"63b375e9-d94c-4e34-882c-be534f9c046b\" (UID: \"63b375e9-d94c-4e34-882c-be534f9c046b\") " Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.568524 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b375e9-d94c-4e34-882c-be534f9c046b-config-volume" (OuterVolumeSpecName: "config-volume") pod "63b375e9-d94c-4e34-882c-be534f9c046b" (UID: "63b375e9-d94c-4e34-882c-be534f9c046b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.576041 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b375e9-d94c-4e34-882c-be534f9c046b-kube-api-access-xc9g9" (OuterVolumeSpecName: "kube-api-access-xc9g9") pod "63b375e9-d94c-4e34-882c-be534f9c046b" (UID: "63b375e9-d94c-4e34-882c-be534f9c046b"). InnerVolumeSpecName "kube-api-access-xc9g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.576376 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63b375e9-d94c-4e34-882c-be534f9c046b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "63b375e9-d94c-4e34-882c-be534f9c046b" (UID: "63b375e9-d94c-4e34-882c-be534f9c046b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.668484 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/63b375e9-d94c-4e34-882c-be534f9c046b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.668526 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/63b375e9-d94c-4e34-882c-be534f9c046b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.668536 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc9g9\" (UniqueName: \"kubernetes.io/projected/63b375e9-d94c-4e34-882c-be534f9c046b-kube-api-access-xc9g9\") on node \"crc\" DevicePath \"\"" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.972898 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" event={"ID":"63b375e9-d94c-4e34-882c-be534f9c046b","Type":"ContainerDied","Data":"ac1f98654a45487066dc62ce78408749a87794b67926a0e35fe1f98b0d8b684f"} Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.972956 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac1f98654a45487066dc62ce78408749a87794b67926a0e35fe1f98b0d8b684f" Dec 03 02:00:04 crc kubenswrapper[4912]: I1203 02:00:04.972924 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj" Dec 03 02:00:05 crc kubenswrapper[4912]: I1203 02:00:05.543833 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq"] Dec 03 02:00:05 crc kubenswrapper[4912]: I1203 02:00:05.557099 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412075-sstpq"] Dec 03 02:00:05 crc kubenswrapper[4912]: I1203 02:00:05.995932 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cz2nd" podUID="244bf195-a521-4253-bb1d-52105460dd83" containerName="registry-server" containerID="cri-o://fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd" gracePeriod=2 Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.564709 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cz2nd" Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.592777 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0038382a-a3be-484a-a163-e1dc6f8cfb21" path="/var/lib/kubelet/pods/0038382a-a3be-484a-a163-e1dc6f8cfb21/volumes" Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.726388 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-catalog-content\") pod \"244bf195-a521-4253-bb1d-52105460dd83\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.726463 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkrrx\" (UniqueName: \"kubernetes.io/projected/244bf195-a521-4253-bb1d-52105460dd83-kube-api-access-pkrrx\") pod \"244bf195-a521-4253-bb1d-52105460dd83\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.726506 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-utilities\") pod \"244bf195-a521-4253-bb1d-52105460dd83\" (UID: \"244bf195-a521-4253-bb1d-52105460dd83\") " Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.730501 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-utilities" (OuterVolumeSpecName: "utilities") pod "244bf195-a521-4253-bb1d-52105460dd83" (UID: "244bf195-a521-4253-bb1d-52105460dd83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.737659 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/244bf195-a521-4253-bb1d-52105460dd83-kube-api-access-pkrrx" (OuterVolumeSpecName: "kube-api-access-pkrrx") pod "244bf195-a521-4253-bb1d-52105460dd83" (UID: "244bf195-a521-4253-bb1d-52105460dd83"). InnerVolumeSpecName "kube-api-access-pkrrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.786687 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "244bf195-a521-4253-bb1d-52105460dd83" (UID: "244bf195-a521-4253-bb1d-52105460dd83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.829626 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.829656 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkrrx\" (UniqueName: \"kubernetes.io/projected/244bf195-a521-4253-bb1d-52105460dd83-kube-api-access-pkrrx\") on node \"crc\" DevicePath \"\"" Dec 03 02:00:06 crc kubenswrapper[4912]: I1203 02:00:06.829668 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/244bf195-a521-4253-bb1d-52105460dd83-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.010832 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cz2nd" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.011444 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cz2nd" event={"ID":"244bf195-a521-4253-bb1d-52105460dd83","Type":"ContainerDied","Data":"fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd"} Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.011509 4912 scope.go:117] "RemoveContainer" containerID="fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.010820 4912 generic.go:334] "Generic (PLEG): container finished" podID="244bf195-a521-4253-bb1d-52105460dd83" containerID="fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd" exitCode=0 Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.011716 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cz2nd" event={"ID":"244bf195-a521-4253-bb1d-52105460dd83","Type":"ContainerDied","Data":"99808ff95106da96d75e62caa54e70836ad5002c8d22f3363bcc67410a34a6f6"} Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.037947 4912 scope.go:117] "RemoveContainer" containerID="4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.056863 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cz2nd"] Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.069591 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cz2nd"] Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.079330 4912 scope.go:117] "RemoveContainer" containerID="1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.130299 4912 scope.go:117] "RemoveContainer" containerID="fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd" Dec 03 02:00:07 crc kubenswrapper[4912]: E1203 02:00:07.130790 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd\": container with ID starting with fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd not found: ID does not exist" containerID="fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.130820 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd"} err="failed to get container status \"fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd\": rpc error: code = NotFound desc = could not find container \"fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd\": container with ID starting with fab27536f6a16ac3806103ffc65ae7976d641aa441c42ec31068b7598fb51fdd not found: ID does not exist" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.130840 4912 scope.go:117] "RemoveContainer" containerID="4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b" Dec 03 02:00:07 crc kubenswrapper[4912]: E1203 02:00:07.131121 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b\": container with ID starting with 4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b not found: ID does not exist" containerID="4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.131141 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b"} err="failed to get container status \"4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b\": rpc error: code = NotFound desc = could not find container \"4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b\": container with ID starting with 4e92cc94130e8fb7c3e21a21f03d54e082f9d8a62e2a6b25b25abbc83603cf3b not found: ID does not exist" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.131154 4912 scope.go:117] "RemoveContainer" containerID="1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288" Dec 03 02:00:07 crc kubenswrapper[4912]: E1203 02:00:07.131371 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288\": container with ID starting with 1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288 not found: ID does not exist" containerID="1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288" Dec 03 02:00:07 crc kubenswrapper[4912]: I1203 02:00:07.131387 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288"} err="failed to get container status \"1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288\": rpc error: code = NotFound desc = could not find container \"1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288\": container with ID starting with 1a19e273ea40adadfc5cadfbdb45fec34b6093042f75ff688d1e9a1a79add288 not found: ID does not exist" Dec 03 02:00:08 crc kubenswrapper[4912]: I1203 02:00:08.586545 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="244bf195-a521-4253-bb1d-52105460dd83" path="/var/lib/kubelet/pods/244bf195-a521-4253-bb1d-52105460dd83/volumes" Dec 03 02:00:09 crc kubenswrapper[4912]: I1203 02:00:09.572656 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:00:09 crc kubenswrapper[4912]: E1203 02:00:09.573138 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:00:21 crc kubenswrapper[4912]: I1203 02:00:21.572292 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:00:21 crc kubenswrapper[4912]: E1203 02:00:21.573100 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:00:28 crc kubenswrapper[4912]: I1203 02:00:28.070902 4912 scope.go:117] "RemoveContainer" containerID="6dba28d35469ad6d40a21cc1257355c24547d53f6c90226cc1f0f0027ee81ff6" Dec 03 02:00:28 crc kubenswrapper[4912]: I1203 02:00:28.141366 4912 scope.go:117] "RemoveContainer" containerID="2e164f6062c79afe348e86fcd5e0dabe2d990d246afd65719452b71c35028c7b" Dec 03 02:00:36 crc kubenswrapper[4912]: I1203 02:00:36.572290 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:00:36 crc kubenswrapper[4912]: E1203 02:00:36.573163 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:00:49 crc kubenswrapper[4912]: I1203 02:00:49.572852 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:00:49 crc kubenswrapper[4912]: E1203 02:00:49.573479 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.158703 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29412121-dkhrq"] Dec 03 02:01:00 crc kubenswrapper[4912]: E1203 02:01:00.160587 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244bf195-a521-4253-bb1d-52105460dd83" containerName="extract-utilities" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.160697 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="244bf195-a521-4253-bb1d-52105460dd83" containerName="extract-utilities" Dec 03 02:01:00 crc kubenswrapper[4912]: E1203 02:01:00.160816 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b375e9-d94c-4e34-882c-be534f9c046b" containerName="collect-profiles" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.160886 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b375e9-d94c-4e34-882c-be534f9c046b" containerName="collect-profiles" Dec 03 02:01:00 crc kubenswrapper[4912]: E1203 02:01:00.160992 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244bf195-a521-4253-bb1d-52105460dd83" containerName="registry-server" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.161071 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="244bf195-a521-4253-bb1d-52105460dd83" containerName="registry-server" Dec 03 02:01:00 crc kubenswrapper[4912]: E1203 02:01:00.161149 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244bf195-a521-4253-bb1d-52105460dd83" containerName="extract-content" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.161220 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="244bf195-a521-4253-bb1d-52105460dd83" containerName="extract-content" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.161578 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="244bf195-a521-4253-bb1d-52105460dd83" containerName="registry-server" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.161686 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="63b375e9-d94c-4e34-882c-be534f9c046b" containerName="collect-profiles" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.162627 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.174728 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412121-dkhrq"] Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.241935 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-combined-ca-bundle\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.242023 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-fernet-keys\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.242097 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-config-data\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.242115 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n44bt\" (UniqueName: \"kubernetes.io/projected/bb6b4855-3d83-4341-942e-de0bc1027bc4-kube-api-access-n44bt\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.345414 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-combined-ca-bundle\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.345936 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-fernet-keys\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.346204 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n44bt\" (UniqueName: \"kubernetes.io/projected/bb6b4855-3d83-4341-942e-de0bc1027bc4-kube-api-access-n44bt\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.346343 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-config-data\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.352492 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-fernet-keys\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.354651 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-config-data\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.356940 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-combined-ca-bundle\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.365594 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n44bt\" (UniqueName: \"kubernetes.io/projected/bb6b4855-3d83-4341-942e-de0bc1027bc4-kube-api-access-n44bt\") pod \"keystone-cron-29412121-dkhrq\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.485802 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:00 crc kubenswrapper[4912]: I1203 02:01:00.572174 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:01:00 crc kubenswrapper[4912]: E1203 02:01:00.572526 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:01:01 crc kubenswrapper[4912]: I1203 02:01:01.053489 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29412121-dkhrq"] Dec 03 02:01:01 crc kubenswrapper[4912]: I1203 02:01:01.694128 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412121-dkhrq" event={"ID":"bb6b4855-3d83-4341-942e-de0bc1027bc4","Type":"ContainerStarted","Data":"67583199463bd8f94a05a74dc8643d2e394316317f7c021526ccd6cc9b725bb0"} Dec 03 02:01:01 crc kubenswrapper[4912]: I1203 02:01:01.694508 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412121-dkhrq" event={"ID":"bb6b4855-3d83-4341-942e-de0bc1027bc4","Type":"ContainerStarted","Data":"26f2b119f531af27eb0257a749e24b7c7be58a78eff1a824abc3e6f10aae6429"} Dec 03 02:01:01 crc kubenswrapper[4912]: I1203 02:01:01.725036 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29412121-dkhrq" podStartSLOduration=1.725013028 podStartE2EDuration="1.725013028s" podCreationTimestamp="2025-12-03 02:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 02:01:01.724395161 +0000 UTC m=+5847.366415791" watchObservedRunningTime="2025-12-03 02:01:01.725013028 +0000 UTC m=+5847.367033588" Dec 03 02:01:03 crc kubenswrapper[4912]: I1203 02:01:03.720690 4912 generic.go:334] "Generic (PLEG): container finished" podID="bb6b4855-3d83-4341-942e-de0bc1027bc4" containerID="67583199463bd8f94a05a74dc8643d2e394316317f7c021526ccd6cc9b725bb0" exitCode=0 Dec 03 02:01:03 crc kubenswrapper[4912]: I1203 02:01:03.720866 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412121-dkhrq" event={"ID":"bb6b4855-3d83-4341-942e-de0bc1027bc4","Type":"ContainerDied","Data":"67583199463bd8f94a05a74dc8643d2e394316317f7c021526ccd6cc9b725bb0"} Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.190582 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.259296 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n44bt\" (UniqueName: \"kubernetes.io/projected/bb6b4855-3d83-4341-942e-de0bc1027bc4-kube-api-access-n44bt\") pod \"bb6b4855-3d83-4341-942e-de0bc1027bc4\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.259624 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-config-data\") pod \"bb6b4855-3d83-4341-942e-de0bc1027bc4\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.259729 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-combined-ca-bundle\") pod \"bb6b4855-3d83-4341-942e-de0bc1027bc4\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.259864 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-fernet-keys\") pod \"bb6b4855-3d83-4341-942e-de0bc1027bc4\" (UID: \"bb6b4855-3d83-4341-942e-de0bc1027bc4\") " Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.267705 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb6b4855-3d83-4341-942e-de0bc1027bc4-kube-api-access-n44bt" (OuterVolumeSpecName: "kube-api-access-n44bt") pod "bb6b4855-3d83-4341-942e-de0bc1027bc4" (UID: "bb6b4855-3d83-4341-942e-de0bc1027bc4"). InnerVolumeSpecName "kube-api-access-n44bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.276617 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bb6b4855-3d83-4341-942e-de0bc1027bc4" (UID: "bb6b4855-3d83-4341-942e-de0bc1027bc4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.326506 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb6b4855-3d83-4341-942e-de0bc1027bc4" (UID: "bb6b4855-3d83-4341-942e-de0bc1027bc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.355931 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-config-data" (OuterVolumeSpecName: "config-data") pod "bb6b4855-3d83-4341-942e-de0bc1027bc4" (UID: "bb6b4855-3d83-4341-942e-de0bc1027bc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.363519 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n44bt\" (UniqueName: \"kubernetes.io/projected/bb6b4855-3d83-4341-942e-de0bc1027bc4-kube-api-access-n44bt\") on node \"crc\" DevicePath \"\"" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.363563 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.363584 4912 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.363600 4912 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bb6b4855-3d83-4341-942e-de0bc1027bc4-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.745909 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29412121-dkhrq" event={"ID":"bb6b4855-3d83-4341-942e-de0bc1027bc4","Type":"ContainerDied","Data":"26f2b119f531af27eb0257a749e24b7c7be58a78eff1a824abc3e6f10aae6429"} Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.746191 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26f2b119f531af27eb0257a749e24b7c7be58a78eff1a824abc3e6f10aae6429" Dec 03 02:01:05 crc kubenswrapper[4912]: I1203 02:01:05.745969 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29412121-dkhrq" Dec 03 02:01:05 crc kubenswrapper[4912]: E1203 02:01:05.873387 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb6b4855_3d83_4341_942e_de0bc1027bc4.slice\": RecentStats: unable to find data in memory cache]" Dec 03 02:01:15 crc kubenswrapper[4912]: I1203 02:01:15.572385 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:01:15 crc kubenswrapper[4912]: E1203 02:01:15.573487 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:01:27 crc kubenswrapper[4912]: I1203 02:01:27.573092 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:01:27 crc kubenswrapper[4912]: E1203 02:01:27.574515 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:01:38 crc kubenswrapper[4912]: I1203 02:01:38.573082 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:01:38 crc kubenswrapper[4912]: E1203 02:01:38.575868 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:01:51 crc kubenswrapper[4912]: I1203 02:01:51.572536 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:01:51 crc kubenswrapper[4912]: E1203 02:01:51.575059 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:02:05 crc kubenswrapper[4912]: I1203 02:02:05.572923 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:02:05 crc kubenswrapper[4912]: E1203 02:02:05.575359 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:02:17 crc kubenswrapper[4912]: I1203 02:02:17.572127 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:02:17 crc kubenswrapper[4912]: E1203 02:02:17.573325 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:02:28 crc kubenswrapper[4912]: I1203 02:02:28.572201 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:02:28 crc kubenswrapper[4912]: E1203 02:02:28.572884 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:02:42 crc kubenswrapper[4912]: I1203 02:02:42.572230 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:02:42 crc kubenswrapper[4912]: E1203 02:02:42.573352 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:02:54 crc kubenswrapper[4912]: I1203 02:02:54.584010 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:02:54 crc kubenswrapper[4912]: E1203 02:02:54.584976 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:03:08 crc kubenswrapper[4912]: I1203 02:03:08.578485 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:03:08 crc kubenswrapper[4912]: E1203 02:03:08.579293 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:03:20 crc kubenswrapper[4912]: I1203 02:03:20.573064 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:03:21 crc kubenswrapper[4912]: I1203 02:03:21.494163 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"4ae78c576371f3feea82d3a850b23c2d1951e6f5b3eb7c72500ce0b9ca508010"} Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.305692 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g68tg"] Dec 03 02:04:21 crc kubenswrapper[4912]: E1203 02:04:21.307078 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb6b4855-3d83-4341-942e-de0bc1027bc4" containerName="keystone-cron" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.307100 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb6b4855-3d83-4341-942e-de0bc1027bc4" containerName="keystone-cron" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.307541 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb6b4855-3d83-4341-942e-de0bc1027bc4" containerName="keystone-cron" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.310265 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.319962 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g68tg"] Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.341749 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsxrh\" (UniqueName: \"kubernetes.io/projected/578ded94-5678-4dec-b920-6047058cb707-kube-api-access-qsxrh\") pod \"redhat-marketplace-g68tg\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.341824 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-catalog-content\") pod \"redhat-marketplace-g68tg\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.341862 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-utilities\") pod \"redhat-marketplace-g68tg\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.442665 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsxrh\" (UniqueName: \"kubernetes.io/projected/578ded94-5678-4dec-b920-6047058cb707-kube-api-access-qsxrh\") pod \"redhat-marketplace-g68tg\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.442756 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-catalog-content\") pod \"redhat-marketplace-g68tg\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.442783 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-utilities\") pod \"redhat-marketplace-g68tg\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.443367 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-catalog-content\") pod \"redhat-marketplace-g68tg\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.443480 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-utilities\") pod \"redhat-marketplace-g68tg\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.462275 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsxrh\" (UniqueName: \"kubernetes.io/projected/578ded94-5678-4dec-b920-6047058cb707-kube-api-access-qsxrh\") pod \"redhat-marketplace-g68tg\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:21 crc kubenswrapper[4912]: I1203 02:04:21.651393 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:22 crc kubenswrapper[4912]: I1203 02:04:22.173671 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g68tg"] Dec 03 02:04:22 crc kubenswrapper[4912]: W1203 02:04:22.185267 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod578ded94_5678_4dec_b920_6047058cb707.slice/crio-ec935c5c5b869f2c293de8f75db4682bf436a3c2ed3466d9b33a12ea291a381b WatchSource:0}: Error finding container ec935c5c5b869f2c293de8f75db4682bf436a3c2ed3466d9b33a12ea291a381b: Status 404 returned error can't find the container with id ec935c5c5b869f2c293de8f75db4682bf436a3c2ed3466d9b33a12ea291a381b Dec 03 02:04:22 crc kubenswrapper[4912]: I1203 02:04:22.666901 4912 generic.go:334] "Generic (PLEG): container finished" podID="578ded94-5678-4dec-b920-6047058cb707" containerID="763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6" exitCode=0 Dec 03 02:04:22 crc kubenswrapper[4912]: I1203 02:04:22.667040 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g68tg" event={"ID":"578ded94-5678-4dec-b920-6047058cb707","Type":"ContainerDied","Data":"763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6"} Dec 03 02:04:22 crc kubenswrapper[4912]: I1203 02:04:22.667178 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g68tg" event={"ID":"578ded94-5678-4dec-b920-6047058cb707","Type":"ContainerStarted","Data":"ec935c5c5b869f2c293de8f75db4682bf436a3c2ed3466d9b33a12ea291a381b"} Dec 03 02:04:22 crc kubenswrapper[4912]: I1203 02:04:22.672323 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 02:04:23 crc kubenswrapper[4912]: I1203 02:04:23.682125 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g68tg" event={"ID":"578ded94-5678-4dec-b920-6047058cb707","Type":"ContainerStarted","Data":"bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26"} Dec 03 02:04:24 crc kubenswrapper[4912]: I1203 02:04:24.698030 4912 generic.go:334] "Generic (PLEG): container finished" podID="578ded94-5678-4dec-b920-6047058cb707" containerID="bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26" exitCode=0 Dec 03 02:04:24 crc kubenswrapper[4912]: I1203 02:04:24.698110 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g68tg" event={"ID":"578ded94-5678-4dec-b920-6047058cb707","Type":"ContainerDied","Data":"bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26"} Dec 03 02:04:25 crc kubenswrapper[4912]: I1203 02:04:25.709752 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g68tg" event={"ID":"578ded94-5678-4dec-b920-6047058cb707","Type":"ContainerStarted","Data":"506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40"} Dec 03 02:04:25 crc kubenswrapper[4912]: I1203 02:04:25.735364 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g68tg" podStartSLOduration=2.272721482 podStartE2EDuration="4.73530322s" podCreationTimestamp="2025-12-03 02:04:21 +0000 UTC" firstStartedPulling="2025-12-03 02:04:22.670704471 +0000 UTC m=+6048.312725061" lastFinishedPulling="2025-12-03 02:04:25.133286199 +0000 UTC m=+6050.775306799" observedRunningTime="2025-12-03 02:04:25.730018619 +0000 UTC m=+6051.372039199" watchObservedRunningTime="2025-12-03 02:04:25.73530322 +0000 UTC m=+6051.377323780" Dec 03 02:04:31 crc kubenswrapper[4912]: I1203 02:04:31.652329 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:31 crc kubenswrapper[4912]: I1203 02:04:31.653222 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:31 crc kubenswrapper[4912]: I1203 02:04:31.731404 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:31 crc kubenswrapper[4912]: I1203 02:04:31.885535 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:32 crc kubenswrapper[4912]: I1203 02:04:32.016528 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g68tg"] Dec 03 02:04:33 crc kubenswrapper[4912]: I1203 02:04:33.814625 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g68tg" podUID="578ded94-5678-4dec-b920-6047058cb707" containerName="registry-server" containerID="cri-o://506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40" gracePeriod=2 Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.411842 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.547977 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-utilities\") pod \"578ded94-5678-4dec-b920-6047058cb707\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.548089 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-catalog-content\") pod \"578ded94-5678-4dec-b920-6047058cb707\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.548158 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsxrh\" (UniqueName: \"kubernetes.io/projected/578ded94-5678-4dec-b920-6047058cb707-kube-api-access-qsxrh\") pod \"578ded94-5678-4dec-b920-6047058cb707\" (UID: \"578ded94-5678-4dec-b920-6047058cb707\") " Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.587934 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/578ded94-5678-4dec-b920-6047058cb707-kube-api-access-qsxrh" (OuterVolumeSpecName: "kube-api-access-qsxrh") pod "578ded94-5678-4dec-b920-6047058cb707" (UID: "578ded94-5678-4dec-b920-6047058cb707"). InnerVolumeSpecName "kube-api-access-qsxrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.608640 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-utilities" (OuterVolumeSpecName: "utilities") pod "578ded94-5678-4dec-b920-6047058cb707" (UID: "578ded94-5678-4dec-b920-6047058cb707"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.636833 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "578ded94-5678-4dec-b920-6047058cb707" (UID: "578ded94-5678-4dec-b920-6047058cb707"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.653822 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.653860 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsxrh\" (UniqueName: \"kubernetes.io/projected/578ded94-5678-4dec-b920-6047058cb707-kube-api-access-qsxrh\") on node \"crc\" DevicePath \"\"" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.653875 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/578ded94-5678-4dec-b920-6047058cb707-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.831670 4912 generic.go:334] "Generic (PLEG): container finished" podID="578ded94-5678-4dec-b920-6047058cb707" containerID="506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40" exitCode=0 Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.831728 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g68tg" event={"ID":"578ded94-5678-4dec-b920-6047058cb707","Type":"ContainerDied","Data":"506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40"} Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.831800 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g68tg" event={"ID":"578ded94-5678-4dec-b920-6047058cb707","Type":"ContainerDied","Data":"ec935c5c5b869f2c293de8f75db4682bf436a3c2ed3466d9b33a12ea291a381b"} Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.831836 4912 scope.go:117] "RemoveContainer" containerID="506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.831753 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g68tg" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.872017 4912 scope.go:117] "RemoveContainer" containerID="bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.892822 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g68tg"] Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.910873 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g68tg"] Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.915224 4912 scope.go:117] "RemoveContainer" containerID="763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.977159 4912 scope.go:117] "RemoveContainer" containerID="506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40" Dec 03 02:04:34 crc kubenswrapper[4912]: E1203 02:04:34.977926 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40\": container with ID starting with 506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40 not found: ID does not exist" containerID="506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.977959 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40"} err="failed to get container status \"506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40\": rpc error: code = NotFound desc = could not find container \"506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40\": container with ID starting with 506b94e0654ca95b947d9085486e77b336482605bba25c2dd563a687e8e53e40 not found: ID does not exist" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.977984 4912 scope.go:117] "RemoveContainer" containerID="bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26" Dec 03 02:04:34 crc kubenswrapper[4912]: E1203 02:04:34.978993 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26\": container with ID starting with bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26 not found: ID does not exist" containerID="bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.979025 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26"} err="failed to get container status \"bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26\": rpc error: code = NotFound desc = could not find container \"bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26\": container with ID starting with bb7bbb2a6277a326b06cd1ed044597f14d4f4e45425e76e9d7e63cca5f36cd26 not found: ID does not exist" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.979046 4912 scope.go:117] "RemoveContainer" containerID="763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6" Dec 03 02:04:34 crc kubenswrapper[4912]: E1203 02:04:34.979539 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6\": container with ID starting with 763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6 not found: ID does not exist" containerID="763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6" Dec 03 02:04:34 crc kubenswrapper[4912]: I1203 02:04:34.979567 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6"} err="failed to get container status \"763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6\": rpc error: code = NotFound desc = could not find container \"763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6\": container with ID starting with 763d6ee76d53e28e2ec7e97f41c0e44bd23f5a9ed4773b8d0fd449b6de2b0ae6 not found: ID does not exist" Dec 03 02:04:36 crc kubenswrapper[4912]: I1203 02:04:36.604036 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="578ded94-5678-4dec-b920-6047058cb707" path="/var/lib/kubelet/pods/578ded94-5678-4dec-b920-6047058cb707/volumes" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.328298 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-74n8m"] Dec 03 02:05:44 crc kubenswrapper[4912]: E1203 02:05:44.329380 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="578ded94-5678-4dec-b920-6047058cb707" containerName="registry-server" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.329396 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="578ded94-5678-4dec-b920-6047058cb707" containerName="registry-server" Dec 03 02:05:44 crc kubenswrapper[4912]: E1203 02:05:44.329418 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="578ded94-5678-4dec-b920-6047058cb707" containerName="extract-content" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.329442 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="578ded94-5678-4dec-b920-6047058cb707" containerName="extract-content" Dec 03 02:05:44 crc kubenswrapper[4912]: E1203 02:05:44.329456 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="578ded94-5678-4dec-b920-6047058cb707" containerName="extract-utilities" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.329462 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="578ded94-5678-4dec-b920-6047058cb707" containerName="extract-utilities" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.329716 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="578ded94-5678-4dec-b920-6047058cb707" containerName="registry-server" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.331265 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.357617 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74n8m"] Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.490559 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-utilities\") pod \"redhat-operators-74n8m\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.490618 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-catalog-content\") pod \"redhat-operators-74n8m\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.490686 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84mhc\" (UniqueName: \"kubernetes.io/projected/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-kube-api-access-84mhc\") pod \"redhat-operators-74n8m\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.592867 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-utilities\") pod \"redhat-operators-74n8m\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.593261 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-catalog-content\") pod \"redhat-operators-74n8m\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.593352 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84mhc\" (UniqueName: \"kubernetes.io/projected/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-kube-api-access-84mhc\") pod \"redhat-operators-74n8m\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.593494 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-utilities\") pod \"redhat-operators-74n8m\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.593714 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-catalog-content\") pod \"redhat-operators-74n8m\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.633852 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84mhc\" (UniqueName: \"kubernetes.io/projected/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-kube-api-access-84mhc\") pod \"redhat-operators-74n8m\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:44 crc kubenswrapper[4912]: I1203 02:05:44.664975 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:45 crc kubenswrapper[4912]: W1203 02:05:45.198366 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec0abb1e_cf7b_481e_b28c_9b7d9b95d4b2.slice/crio-49d232404150a446c3860de91ef41b804c3eeab401561d25f08198934cdd302d WatchSource:0}: Error finding container 49d232404150a446c3860de91ef41b804c3eeab401561d25f08198934cdd302d: Status 404 returned error can't find the container with id 49d232404150a446c3860de91ef41b804c3eeab401561d25f08198934cdd302d Dec 03 02:05:45 crc kubenswrapper[4912]: I1203 02:05:45.199350 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74n8m"] Dec 03 02:05:46 crc kubenswrapper[4912]: I1203 02:05:46.047807 4912 generic.go:334] "Generic (PLEG): container finished" podID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerID="879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd" exitCode=0 Dec 03 02:05:46 crc kubenswrapper[4912]: I1203 02:05:46.047855 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74n8m" event={"ID":"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2","Type":"ContainerDied","Data":"879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd"} Dec 03 02:05:46 crc kubenswrapper[4912]: I1203 02:05:46.048246 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74n8m" event={"ID":"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2","Type":"ContainerStarted","Data":"49d232404150a446c3860de91ef41b804c3eeab401561d25f08198934cdd302d"} Dec 03 02:05:47 crc kubenswrapper[4912]: I1203 02:05:47.058771 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74n8m" event={"ID":"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2","Type":"ContainerStarted","Data":"f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc"} Dec 03 02:05:48 crc kubenswrapper[4912]: I1203 02:05:48.078539 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:05:48 crc kubenswrapper[4912]: I1203 02:05:48.079226 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:05:50 crc kubenswrapper[4912]: I1203 02:05:50.097709 4912 generic.go:334] "Generic (PLEG): container finished" podID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerID="f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc" exitCode=0 Dec 03 02:05:50 crc kubenswrapper[4912]: I1203 02:05:50.097782 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74n8m" event={"ID":"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2","Type":"ContainerDied","Data":"f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc"} Dec 03 02:05:51 crc kubenswrapper[4912]: I1203 02:05:51.110555 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74n8m" event={"ID":"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2","Type":"ContainerStarted","Data":"dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133"} Dec 03 02:05:51 crc kubenswrapper[4912]: I1203 02:05:51.140098 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-74n8m" podStartSLOduration=2.706062187 podStartE2EDuration="7.140075161s" podCreationTimestamp="2025-12-03 02:05:44 +0000 UTC" firstStartedPulling="2025-12-03 02:05:46.050086214 +0000 UTC m=+6131.692106774" lastFinishedPulling="2025-12-03 02:05:50.484099188 +0000 UTC m=+6136.126119748" observedRunningTime="2025-12-03 02:05:51.134566223 +0000 UTC m=+6136.776586783" watchObservedRunningTime="2025-12-03 02:05:51.140075161 +0000 UTC m=+6136.782095721" Dec 03 02:05:54 crc kubenswrapper[4912]: I1203 02:05:54.666197 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:54 crc kubenswrapper[4912]: I1203 02:05:54.666836 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:05:55 crc kubenswrapper[4912]: I1203 02:05:55.740373 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-74n8m" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerName="registry-server" probeResult="failure" output=< Dec 03 02:05:55 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 02:05:55 crc kubenswrapper[4912]: > Dec 03 02:06:04 crc kubenswrapper[4912]: I1203 02:06:04.735288 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:06:04 crc kubenswrapper[4912]: I1203 02:06:04.803951 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:06:04 crc kubenswrapper[4912]: I1203 02:06:04.987396 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74n8m"] Dec 03 02:06:06 crc kubenswrapper[4912]: I1203 02:06:06.283943 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-74n8m" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerName="registry-server" containerID="cri-o://dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133" gracePeriod=2 Dec 03 02:06:06 crc kubenswrapper[4912]: I1203 02:06:06.863100 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.054062 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-utilities\") pod \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.054288 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84mhc\" (UniqueName: \"kubernetes.io/projected/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-kube-api-access-84mhc\") pod \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.054345 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-catalog-content\") pod \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\" (UID: \"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2\") " Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.055557 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-utilities" (OuterVolumeSpecName: "utilities") pod "ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" (UID: "ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.067766 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-kube-api-access-84mhc" (OuterVolumeSpecName: "kube-api-access-84mhc") pod "ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" (UID: "ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2"). InnerVolumeSpecName "kube-api-access-84mhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.157490 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.157525 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84mhc\" (UniqueName: \"kubernetes.io/projected/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-kube-api-access-84mhc\") on node \"crc\" DevicePath \"\"" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.237601 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" (UID: "ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.260213 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.332763 4912 generic.go:334] "Generic (PLEG): container finished" podID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerID="dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133" exitCode=0 Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.333035 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74n8m" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.333622 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74n8m" event={"ID":"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2","Type":"ContainerDied","Data":"dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133"} Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.333708 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74n8m" event={"ID":"ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2","Type":"ContainerDied","Data":"49d232404150a446c3860de91ef41b804c3eeab401561d25f08198934cdd302d"} Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.333727 4912 scope.go:117] "RemoveContainer" containerID="dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.356161 4912 scope.go:117] "RemoveContainer" containerID="f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.375488 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74n8m"] Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.385599 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-74n8m"] Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.399736 4912 scope.go:117] "RemoveContainer" containerID="879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.451033 4912 scope.go:117] "RemoveContainer" containerID="dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133" Dec 03 02:06:07 crc kubenswrapper[4912]: E1203 02:06:07.451955 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133\": container with ID starting with dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133 not found: ID does not exist" containerID="dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.451989 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133"} err="failed to get container status \"dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133\": rpc error: code = NotFound desc = could not find container \"dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133\": container with ID starting with dcbda52a8ecffc3783e4664f612847ca57c805de55e1c067fba629059badc133 not found: ID does not exist" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.452011 4912 scope.go:117] "RemoveContainer" containerID="f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc" Dec 03 02:06:07 crc kubenswrapper[4912]: E1203 02:06:07.452552 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc\": container with ID starting with f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc not found: ID does not exist" containerID="f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.452576 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc"} err="failed to get container status \"f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc\": rpc error: code = NotFound desc = could not find container \"f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc\": container with ID starting with f381fa5d637fa272b3d1bb24d352b043feb7d0ebbb3cbcf5b553ea3546fdf2dc not found: ID does not exist" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.452590 4912 scope.go:117] "RemoveContainer" containerID="879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd" Dec 03 02:06:07 crc kubenswrapper[4912]: E1203 02:06:07.453090 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd\": container with ID starting with 879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd not found: ID does not exist" containerID="879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd" Dec 03 02:06:07 crc kubenswrapper[4912]: I1203 02:06:07.453111 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd"} err="failed to get container status \"879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd\": rpc error: code = NotFound desc = could not find container \"879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd\": container with ID starting with 879544ea3115162f7d12ac1b07c5941c40f9b74f3882af811a58401a3b8810cd not found: ID does not exist" Dec 03 02:06:08 crc kubenswrapper[4912]: I1203 02:06:08.586831 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" path="/var/lib/kubelet/pods/ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2/volumes" Dec 03 02:06:18 crc kubenswrapper[4912]: I1203 02:06:18.078234 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:06:18 crc kubenswrapper[4912]: I1203 02:06:18.078857 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:06:48 crc kubenswrapper[4912]: I1203 02:06:48.077635 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:06:48 crc kubenswrapper[4912]: I1203 02:06:48.078303 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:06:48 crc kubenswrapper[4912]: I1203 02:06:48.078713 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:06:48 crc kubenswrapper[4912]: I1203 02:06:48.079898 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ae78c576371f3feea82d3a850b23c2d1951e6f5b3eb7c72500ce0b9ca508010"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:06:48 crc kubenswrapper[4912]: I1203 02:06:48.079999 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://4ae78c576371f3feea82d3a850b23c2d1951e6f5b3eb7c72500ce0b9ca508010" gracePeriod=600 Dec 03 02:06:48 crc kubenswrapper[4912]: I1203 02:06:48.913601 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="4ae78c576371f3feea82d3a850b23c2d1951e6f5b3eb7c72500ce0b9ca508010" exitCode=0 Dec 03 02:06:48 crc kubenswrapper[4912]: I1203 02:06:48.913754 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"4ae78c576371f3feea82d3a850b23c2d1951e6f5b3eb7c72500ce0b9ca508010"} Dec 03 02:06:48 crc kubenswrapper[4912]: I1203 02:06:48.914061 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da"} Dec 03 02:06:48 crc kubenswrapper[4912]: I1203 02:06:48.914115 4912 scope.go:117] "RemoveContainer" containerID="67612681ca5d260f47896b961055e6a8317d1aa0513657452ea36665583d0879" Dec 03 02:08:48 crc kubenswrapper[4912]: I1203 02:08:48.077987 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:08:48 crc kubenswrapper[4912]: I1203 02:08:48.078736 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.660626 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qhgkz"] Dec 03 02:08:58 crc kubenswrapper[4912]: E1203 02:08:58.661788 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerName="extract-utilities" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.661840 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerName="extract-utilities" Dec 03 02:08:58 crc kubenswrapper[4912]: E1203 02:08:58.661858 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerName="registry-server" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.661865 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerName="registry-server" Dec 03 02:08:58 crc kubenswrapper[4912]: E1203 02:08:58.661907 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerName="extract-content" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.661913 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerName="extract-content" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.662142 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0abb1e-cf7b-481e-b28c-9b7d9b95d4b2" containerName="registry-server" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.663890 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.684363 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qhgkz"] Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.810891 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svznx\" (UniqueName: \"kubernetes.io/projected/abff09ed-57b7-4807-ad69-e50102008613-kube-api-access-svznx\") pod \"certified-operators-qhgkz\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.810965 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-catalog-content\") pod \"certified-operators-qhgkz\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.811001 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-utilities\") pod \"certified-operators-qhgkz\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.912887 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svznx\" (UniqueName: \"kubernetes.io/projected/abff09ed-57b7-4807-ad69-e50102008613-kube-api-access-svznx\") pod \"certified-operators-qhgkz\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.913152 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-catalog-content\") pod \"certified-operators-qhgkz\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.913238 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-utilities\") pod \"certified-operators-qhgkz\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.913706 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-utilities\") pod \"certified-operators-qhgkz\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.913779 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-catalog-content\") pod \"certified-operators-qhgkz\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.937191 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svznx\" (UniqueName: \"kubernetes.io/projected/abff09ed-57b7-4807-ad69-e50102008613-kube-api-access-svznx\") pod \"certified-operators-qhgkz\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:58 crc kubenswrapper[4912]: I1203 02:08:58.996878 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:08:59 crc kubenswrapper[4912]: I1203 02:08:59.531624 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qhgkz"] Dec 03 02:09:00 crc kubenswrapper[4912]: I1203 02:09:00.056531 4912 generic.go:334] "Generic (PLEG): container finished" podID="abff09ed-57b7-4807-ad69-e50102008613" containerID="9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70" exitCode=0 Dec 03 02:09:00 crc kubenswrapper[4912]: I1203 02:09:00.056607 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhgkz" event={"ID":"abff09ed-57b7-4807-ad69-e50102008613","Type":"ContainerDied","Data":"9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70"} Dec 03 02:09:00 crc kubenswrapper[4912]: I1203 02:09:00.056785 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhgkz" event={"ID":"abff09ed-57b7-4807-ad69-e50102008613","Type":"ContainerStarted","Data":"6d55ac4e3f7d262fc301583f6e842dba8eb4f112c093852b10ccfe77985b0ae1"} Dec 03 02:09:01 crc kubenswrapper[4912]: I1203 02:09:01.068893 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhgkz" event={"ID":"abff09ed-57b7-4807-ad69-e50102008613","Type":"ContainerStarted","Data":"750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af"} Dec 03 02:09:02 crc kubenswrapper[4912]: I1203 02:09:02.081791 4912 generic.go:334] "Generic (PLEG): container finished" podID="abff09ed-57b7-4807-ad69-e50102008613" containerID="750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af" exitCode=0 Dec 03 02:09:02 crc kubenswrapper[4912]: I1203 02:09:02.081928 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhgkz" event={"ID":"abff09ed-57b7-4807-ad69-e50102008613","Type":"ContainerDied","Data":"750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af"} Dec 03 02:09:03 crc kubenswrapper[4912]: I1203 02:09:03.096689 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhgkz" event={"ID":"abff09ed-57b7-4807-ad69-e50102008613","Type":"ContainerStarted","Data":"88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee"} Dec 03 02:09:03 crc kubenswrapper[4912]: I1203 02:09:03.118079 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qhgkz" podStartSLOduration=2.632893455 podStartE2EDuration="5.118055023s" podCreationTimestamp="2025-12-03 02:08:58 +0000 UTC" firstStartedPulling="2025-12-03 02:09:00.059467883 +0000 UTC m=+6325.701488453" lastFinishedPulling="2025-12-03 02:09:02.544629461 +0000 UTC m=+6328.186650021" observedRunningTime="2025-12-03 02:09:03.114828196 +0000 UTC m=+6328.756848826" watchObservedRunningTime="2025-12-03 02:09:03.118055023 +0000 UTC m=+6328.760075623" Dec 03 02:09:08 crc kubenswrapper[4912]: I1203 02:09:08.998241 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:09:09 crc kubenswrapper[4912]: I1203 02:09:08.999695 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:09:09 crc kubenswrapper[4912]: I1203 02:09:09.095455 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:09:09 crc kubenswrapper[4912]: I1203 02:09:09.275869 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:09:09 crc kubenswrapper[4912]: I1203 02:09:09.355050 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qhgkz"] Dec 03 02:09:11 crc kubenswrapper[4912]: I1203 02:09:11.220325 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qhgkz" podUID="abff09ed-57b7-4807-ad69-e50102008613" containerName="registry-server" containerID="cri-o://88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee" gracePeriod=2 Dec 03 02:09:11 crc kubenswrapper[4912]: I1203 02:09:11.970328 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.134735 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-catalog-content\") pod \"abff09ed-57b7-4807-ad69-e50102008613\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.134996 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svznx\" (UniqueName: \"kubernetes.io/projected/abff09ed-57b7-4807-ad69-e50102008613-kube-api-access-svznx\") pod \"abff09ed-57b7-4807-ad69-e50102008613\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.135084 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-utilities\") pod \"abff09ed-57b7-4807-ad69-e50102008613\" (UID: \"abff09ed-57b7-4807-ad69-e50102008613\") " Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.137136 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-utilities" (OuterVolumeSpecName: "utilities") pod "abff09ed-57b7-4807-ad69-e50102008613" (UID: "abff09ed-57b7-4807-ad69-e50102008613"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.140047 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abff09ed-57b7-4807-ad69-e50102008613-kube-api-access-svznx" (OuterVolumeSpecName: "kube-api-access-svznx") pod "abff09ed-57b7-4807-ad69-e50102008613" (UID: "abff09ed-57b7-4807-ad69-e50102008613"). InnerVolumeSpecName "kube-api-access-svznx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.203950 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abff09ed-57b7-4807-ad69-e50102008613" (UID: "abff09ed-57b7-4807-ad69-e50102008613"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.238146 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svznx\" (UniqueName: \"kubernetes.io/projected/abff09ed-57b7-4807-ad69-e50102008613-kube-api-access-svznx\") on node \"crc\" DevicePath \"\"" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.238190 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.238202 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abff09ed-57b7-4807-ad69-e50102008613-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.243722 4912 generic.go:334] "Generic (PLEG): container finished" podID="abff09ed-57b7-4807-ad69-e50102008613" containerID="88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee" exitCode=0 Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.243765 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhgkz" event={"ID":"abff09ed-57b7-4807-ad69-e50102008613","Type":"ContainerDied","Data":"88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee"} Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.243794 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qhgkz" event={"ID":"abff09ed-57b7-4807-ad69-e50102008613","Type":"ContainerDied","Data":"6d55ac4e3f7d262fc301583f6e842dba8eb4f112c093852b10ccfe77985b0ae1"} Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.243811 4912 scope.go:117] "RemoveContainer" containerID="88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.243838 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qhgkz" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.282546 4912 scope.go:117] "RemoveContainer" containerID="750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.310669 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qhgkz"] Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.310904 4912 scope.go:117] "RemoveContainer" containerID="9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.326013 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qhgkz"] Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.357769 4912 scope.go:117] "RemoveContainer" containerID="88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee" Dec 03 02:09:12 crc kubenswrapper[4912]: E1203 02:09:12.358224 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee\": container with ID starting with 88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee not found: ID does not exist" containerID="88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.358282 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee"} err="failed to get container status \"88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee\": rpc error: code = NotFound desc = could not find container \"88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee\": container with ID starting with 88f0e9a045bd9ecac74bc27d129c2cba7ed7b3a8c17fc735ca2e32660e737eee not found: ID does not exist" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.358321 4912 scope.go:117] "RemoveContainer" containerID="750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af" Dec 03 02:09:12 crc kubenswrapper[4912]: E1203 02:09:12.359202 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af\": container with ID starting with 750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af not found: ID does not exist" containerID="750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.359231 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af"} err="failed to get container status \"750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af\": rpc error: code = NotFound desc = could not find container \"750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af\": container with ID starting with 750f5f6b79ae7ce0ffa2b8a8286536304e758b4f26af931f41c92c654a1388af not found: ID does not exist" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.359252 4912 scope.go:117] "RemoveContainer" containerID="9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70" Dec 03 02:09:12 crc kubenswrapper[4912]: E1203 02:09:12.359554 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70\": container with ID starting with 9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70 not found: ID does not exist" containerID="9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.359582 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70"} err="failed to get container status \"9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70\": rpc error: code = NotFound desc = could not find container \"9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70\": container with ID starting with 9bbce2ae8543ee70dad1f6fb97de5178de7f1776880ead67cfc9c2020c1cfa70 not found: ID does not exist" Dec 03 02:09:12 crc kubenswrapper[4912]: I1203 02:09:12.591784 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abff09ed-57b7-4807-ad69-e50102008613" path="/var/lib/kubelet/pods/abff09ed-57b7-4807-ad69-e50102008613/volumes" Dec 03 02:09:18 crc kubenswrapper[4912]: I1203 02:09:18.077667 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:09:18 crc kubenswrapper[4912]: I1203 02:09:18.078249 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:09:48 crc kubenswrapper[4912]: I1203 02:09:48.077891 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:09:48 crc kubenswrapper[4912]: I1203 02:09:48.078352 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:09:48 crc kubenswrapper[4912]: I1203 02:09:48.078398 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:09:48 crc kubenswrapper[4912]: I1203 02:09:48.079390 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:09:48 crc kubenswrapper[4912]: I1203 02:09:48.079489 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" gracePeriod=600 Dec 03 02:09:48 crc kubenswrapper[4912]: E1203 02:09:48.207597 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:09:48 crc kubenswrapper[4912]: I1203 02:09:48.769200 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" exitCode=0 Dec 03 02:09:48 crc kubenswrapper[4912]: I1203 02:09:48.769252 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da"} Dec 03 02:09:48 crc kubenswrapper[4912]: I1203 02:09:48.769293 4912 scope.go:117] "RemoveContainer" containerID="4ae78c576371f3feea82d3a850b23c2d1951e6f5b3eb7c72500ce0b9ca508010" Dec 03 02:09:48 crc kubenswrapper[4912]: I1203 02:09:48.770173 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:09:48 crc kubenswrapper[4912]: E1203 02:09:48.770491 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:09:59 crc kubenswrapper[4912]: I1203 02:09:59.572515 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:09:59 crc kubenswrapper[4912]: E1203 02:09:59.573607 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.020517 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t2h4m"] Dec 03 02:10:01 crc kubenswrapper[4912]: E1203 02:10:01.021925 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abff09ed-57b7-4807-ad69-e50102008613" containerName="extract-utilities" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.021959 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="abff09ed-57b7-4807-ad69-e50102008613" containerName="extract-utilities" Dec 03 02:10:01 crc kubenswrapper[4912]: E1203 02:10:01.022055 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abff09ed-57b7-4807-ad69-e50102008613" containerName="registry-server" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.022111 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="abff09ed-57b7-4807-ad69-e50102008613" containerName="registry-server" Dec 03 02:10:01 crc kubenswrapper[4912]: E1203 02:10:01.022150 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abff09ed-57b7-4807-ad69-e50102008613" containerName="extract-content" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.022204 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="abff09ed-57b7-4807-ad69-e50102008613" containerName="extract-content" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.022789 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="abff09ed-57b7-4807-ad69-e50102008613" containerName="registry-server" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.025813 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.032632 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t2h4m"] Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.067993 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-utilities\") pod \"community-operators-t2h4m\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.068155 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n72m\" (UniqueName: \"kubernetes.io/projected/bead5cdc-d271-46e4-904f-2e7f8720d92b-kube-api-access-2n72m\") pod \"community-operators-t2h4m\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.068467 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-catalog-content\") pod \"community-operators-t2h4m\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.171587 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-catalog-content\") pod \"community-operators-t2h4m\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.172181 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-utilities\") pod \"community-operators-t2h4m\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.172310 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-catalog-content\") pod \"community-operators-t2h4m\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.172634 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n72m\" (UniqueName: \"kubernetes.io/projected/bead5cdc-d271-46e4-904f-2e7f8720d92b-kube-api-access-2n72m\") pod \"community-operators-t2h4m\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.172678 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-utilities\") pod \"community-operators-t2h4m\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.192469 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n72m\" (UniqueName: \"kubernetes.io/projected/bead5cdc-d271-46e4-904f-2e7f8720d92b-kube-api-access-2n72m\") pod \"community-operators-t2h4m\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.386378 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.949922 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t2h4m"] Dec 03 02:10:01 crc kubenswrapper[4912]: I1203 02:10:01.987916 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2h4m" event={"ID":"bead5cdc-d271-46e4-904f-2e7f8720d92b","Type":"ContainerStarted","Data":"b9b3f7904eb0c64bcc85a36a89035233428ce90d58abdd6cf84eb52ff2f1fa2e"} Dec 03 02:10:02 crc kubenswrapper[4912]: I1203 02:10:02.998372 4912 generic.go:334] "Generic (PLEG): container finished" podID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerID="b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91" exitCode=0 Dec 03 02:10:02 crc kubenswrapper[4912]: I1203 02:10:02.998476 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2h4m" event={"ID":"bead5cdc-d271-46e4-904f-2e7f8720d92b","Type":"ContainerDied","Data":"b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91"} Dec 03 02:10:03 crc kubenswrapper[4912]: I1203 02:10:03.000787 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 02:10:05 crc kubenswrapper[4912]: I1203 02:10:05.023641 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2h4m" event={"ID":"bead5cdc-d271-46e4-904f-2e7f8720d92b","Type":"ContainerStarted","Data":"ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be"} Dec 03 02:10:06 crc kubenswrapper[4912]: I1203 02:10:06.048701 4912 generic.go:334] "Generic (PLEG): container finished" podID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerID="ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be" exitCode=0 Dec 03 02:10:06 crc kubenswrapper[4912]: I1203 02:10:06.049154 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2h4m" event={"ID":"bead5cdc-d271-46e4-904f-2e7f8720d92b","Type":"ContainerDied","Data":"ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be"} Dec 03 02:10:07 crc kubenswrapper[4912]: I1203 02:10:07.067548 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2h4m" event={"ID":"bead5cdc-d271-46e4-904f-2e7f8720d92b","Type":"ContainerStarted","Data":"dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614"} Dec 03 02:10:07 crc kubenswrapper[4912]: I1203 02:10:07.099138 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t2h4m" podStartSLOduration=3.625254689 podStartE2EDuration="7.099119314s" podCreationTimestamp="2025-12-03 02:10:00 +0000 UTC" firstStartedPulling="2025-12-03 02:10:03.000531776 +0000 UTC m=+6388.642552336" lastFinishedPulling="2025-12-03 02:10:06.474396401 +0000 UTC m=+6392.116416961" observedRunningTime="2025-12-03 02:10:07.087914422 +0000 UTC m=+6392.729934982" watchObservedRunningTime="2025-12-03 02:10:07.099119314 +0000 UTC m=+6392.741139874" Dec 03 02:10:11 crc kubenswrapper[4912]: I1203 02:10:11.387270 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:11 crc kubenswrapper[4912]: I1203 02:10:11.387809 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:11 crc kubenswrapper[4912]: I1203 02:10:11.451343 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:12 crc kubenswrapper[4912]: I1203 02:10:12.198255 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:12 crc kubenswrapper[4912]: I1203 02:10:12.266283 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t2h4m"] Dec 03 02:10:14 crc kubenswrapper[4912]: I1203 02:10:14.151645 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t2h4m" podUID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerName="registry-server" containerID="cri-o://dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614" gracePeriod=2 Dec 03 02:10:14 crc kubenswrapper[4912]: I1203 02:10:14.627014 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:10:14 crc kubenswrapper[4912]: E1203 02:10:14.628019 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:10:14 crc kubenswrapper[4912]: I1203 02:10:14.788939 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:14 crc kubenswrapper[4912]: I1203 02:10:14.957147 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-utilities\") pod \"bead5cdc-d271-46e4-904f-2e7f8720d92b\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " Dec 03 02:10:14 crc kubenswrapper[4912]: I1203 02:10:14.957340 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n72m\" (UniqueName: \"kubernetes.io/projected/bead5cdc-d271-46e4-904f-2e7f8720d92b-kube-api-access-2n72m\") pod \"bead5cdc-d271-46e4-904f-2e7f8720d92b\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " Dec 03 02:10:14 crc kubenswrapper[4912]: I1203 02:10:14.957374 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-catalog-content\") pod \"bead5cdc-d271-46e4-904f-2e7f8720d92b\" (UID: \"bead5cdc-d271-46e4-904f-2e7f8720d92b\") " Dec 03 02:10:14 crc kubenswrapper[4912]: I1203 02:10:14.959499 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-utilities" (OuterVolumeSpecName: "utilities") pod "bead5cdc-d271-46e4-904f-2e7f8720d92b" (UID: "bead5cdc-d271-46e4-904f-2e7f8720d92b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:10:14 crc kubenswrapper[4912]: I1203 02:10:14.967811 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bead5cdc-d271-46e4-904f-2e7f8720d92b-kube-api-access-2n72m" (OuterVolumeSpecName: "kube-api-access-2n72m") pod "bead5cdc-d271-46e4-904f-2e7f8720d92b" (UID: "bead5cdc-d271-46e4-904f-2e7f8720d92b"). InnerVolumeSpecName "kube-api-access-2n72m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.026760 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bead5cdc-d271-46e4-904f-2e7f8720d92b" (UID: "bead5cdc-d271-46e4-904f-2e7f8720d92b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.059968 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n72m\" (UniqueName: \"kubernetes.io/projected/bead5cdc-d271-46e4-904f-2e7f8720d92b-kube-api-access-2n72m\") on node \"crc\" DevicePath \"\"" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.060007 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.060022 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bead5cdc-d271-46e4-904f-2e7f8720d92b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.167977 4912 generic.go:334] "Generic (PLEG): container finished" podID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerID="dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614" exitCode=0 Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.168035 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2h4m" event={"ID":"bead5cdc-d271-46e4-904f-2e7f8720d92b","Type":"ContainerDied","Data":"dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614"} Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.168071 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t2h4m" event={"ID":"bead5cdc-d271-46e4-904f-2e7f8720d92b","Type":"ContainerDied","Data":"b9b3f7904eb0c64bcc85a36a89035233428ce90d58abdd6cf84eb52ff2f1fa2e"} Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.168094 4912 scope.go:117] "RemoveContainer" containerID="dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.168115 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t2h4m" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.226330 4912 scope.go:117] "RemoveContainer" containerID="ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.236707 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t2h4m"] Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.260723 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t2h4m"] Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.261823 4912 scope.go:117] "RemoveContainer" containerID="b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.323352 4912 scope.go:117] "RemoveContainer" containerID="dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614" Dec 03 02:10:15 crc kubenswrapper[4912]: E1203 02:10:15.323913 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614\": container with ID starting with dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614 not found: ID does not exist" containerID="dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.323964 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614"} err="failed to get container status \"dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614\": rpc error: code = NotFound desc = could not find container \"dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614\": container with ID starting with dae08e301425457e407d610b4f14a7a37d56ccdbdb4a37f985cf070faff07614 not found: ID does not exist" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.323996 4912 scope.go:117] "RemoveContainer" containerID="ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be" Dec 03 02:10:15 crc kubenswrapper[4912]: E1203 02:10:15.324524 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be\": container with ID starting with ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be not found: ID does not exist" containerID="ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.324553 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be"} err="failed to get container status \"ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be\": rpc error: code = NotFound desc = could not find container \"ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be\": container with ID starting with ceddbe764e3e80ad7149ced8e2fa10107a872bfec2c1046c37ff74c5a20548be not found: ID does not exist" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.324571 4912 scope.go:117] "RemoveContainer" containerID="b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91" Dec 03 02:10:15 crc kubenswrapper[4912]: E1203 02:10:15.324950 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91\": container with ID starting with b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91 not found: ID does not exist" containerID="b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91" Dec 03 02:10:15 crc kubenswrapper[4912]: I1203 02:10:15.324980 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91"} err="failed to get container status \"b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91\": rpc error: code = NotFound desc = could not find container \"b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91\": container with ID starting with b3ab456640191eb4d44770fac6b0230d18c84e558a8811bb0f934f03de723f91 not found: ID does not exist" Dec 03 02:10:16 crc kubenswrapper[4912]: I1203 02:10:16.593200 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bead5cdc-d271-46e4-904f-2e7f8720d92b" path="/var/lib/kubelet/pods/bead5cdc-d271-46e4-904f-2e7f8720d92b/volumes" Dec 03 02:10:27 crc kubenswrapper[4912]: I1203 02:10:27.573024 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:10:27 crc kubenswrapper[4912]: E1203 02:10:27.574146 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:10:41 crc kubenswrapper[4912]: I1203 02:10:41.572236 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:10:41 crc kubenswrapper[4912]: E1203 02:10:41.573296 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:10:53 crc kubenswrapper[4912]: I1203 02:10:53.572274 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:10:53 crc kubenswrapper[4912]: E1203 02:10:53.573643 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:11:05 crc kubenswrapper[4912]: I1203 02:11:05.572257 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:11:05 crc kubenswrapper[4912]: E1203 02:11:05.573259 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:11:18 crc kubenswrapper[4912]: I1203 02:11:18.572596 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:11:18 crc kubenswrapper[4912]: E1203 02:11:18.573651 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:11:31 crc kubenswrapper[4912]: I1203 02:11:31.572165 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:11:31 crc kubenswrapper[4912]: E1203 02:11:31.573099 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:11:45 crc kubenswrapper[4912]: I1203 02:11:45.571715 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:11:45 crc kubenswrapper[4912]: E1203 02:11:45.572761 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:11:57 crc kubenswrapper[4912]: I1203 02:11:57.573121 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:11:57 crc kubenswrapper[4912]: E1203 02:11:57.574054 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:12:09 crc kubenswrapper[4912]: I1203 02:12:09.571510 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:12:09 crc kubenswrapper[4912]: E1203 02:12:09.572268 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:12:21 crc kubenswrapper[4912]: I1203 02:12:21.572747 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:12:21 crc kubenswrapper[4912]: E1203 02:12:21.573563 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:12:32 crc kubenswrapper[4912]: I1203 02:12:32.573248 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:12:32 crc kubenswrapper[4912]: E1203 02:12:32.574303 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:12:37 crc kubenswrapper[4912]: E1203 02:12:37.448659 4912 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.20:37308->38.102.83.20:41775: write tcp 38.102.83.20:37308->38.102.83.20:41775: write: broken pipe Dec 03 02:12:46 crc kubenswrapper[4912]: I1203 02:12:46.573730 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:12:46 crc kubenswrapper[4912]: E1203 02:12:46.574975 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:13:01 crc kubenswrapper[4912]: I1203 02:13:01.571924 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:13:01 crc kubenswrapper[4912]: E1203 02:13:01.572746 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:13:12 crc kubenswrapper[4912]: I1203 02:13:12.572371 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:13:12 crc kubenswrapper[4912]: E1203 02:13:12.573381 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:13:25 crc kubenswrapper[4912]: I1203 02:13:25.573024 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:13:25 crc kubenswrapper[4912]: E1203 02:13:25.574266 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:13:38 crc kubenswrapper[4912]: I1203 02:13:38.572900 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:13:38 crc kubenswrapper[4912]: E1203 02:13:38.574312 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.026766 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 02:13:39 crc kubenswrapper[4912]: E1203 02:13:39.027836 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerName="registry-server" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.027866 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerName="registry-server" Dec 03 02:13:39 crc kubenswrapper[4912]: E1203 02:13:39.027901 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerName="extract-utilities" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.027914 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerName="extract-utilities" Dec 03 02:13:39 crc kubenswrapper[4912]: E1203 02:13:39.027980 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerName="extract-content" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.027993 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerName="extract-content" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.028388 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="bead5cdc-d271-46e4-904f-2e7f8720d92b" containerName="registry-server" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.029811 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.034179 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.034498 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.034691 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9xb2x" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.036410 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.046685 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw9b2\" (UniqueName: \"kubernetes.io/projected/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-kube-api-access-zw9b2\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.046779 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.046815 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.046893 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-config-data\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.046925 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.046976 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.047004 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.047141 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.047271 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.068777 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.148933 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.149055 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.149115 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw9b2\" (UniqueName: \"kubernetes.io/projected/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-kube-api-access-zw9b2\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.149142 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.149165 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.149208 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-config-data\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.149888 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.150146 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.150340 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-config-data\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.150386 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.150403 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.159134 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.159770 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.159863 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.160842 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.161068 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.172966 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.200205 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw9b2\" (UniqueName: \"kubernetes.io/projected/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-kube-api-access-zw9b2\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.256710 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.361978 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 02:13:39 crc kubenswrapper[4912]: I1203 02:13:39.830154 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 02:13:40 crc kubenswrapper[4912]: I1203 02:13:40.733577 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d5f2ea7f-cdc9-4573-8b16-6561c761f91f","Type":"ContainerStarted","Data":"3dc0ad0e970fcf78de2548b555d30324f15505274865780bdafb05a564b5db28"} Dec 03 02:13:51 crc kubenswrapper[4912]: I1203 02:13:51.572328 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:13:51 crc kubenswrapper[4912]: E1203 02:13:51.573180 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:14:03 crc kubenswrapper[4912]: I1203 02:14:03.571672 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:14:03 crc kubenswrapper[4912]: E1203 02:14:03.572447 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:14:16 crc kubenswrapper[4912]: I1203 02:14:16.572996 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:14:16 crc kubenswrapper[4912]: E1203 02:14:16.573914 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:14:18 crc kubenswrapper[4912]: E1203 02:14:18.607338 4912 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 03 02:14:18 crc kubenswrapper[4912]: E1203 02:14:18.608415 4912 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zw9b2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(d5f2ea7f-cdc9-4573-8b16-6561c761f91f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 02:14:18 crc kubenswrapper[4912]: E1203 02:14:18.609775 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="d5f2ea7f-cdc9-4573-8b16-6561c761f91f" Dec 03 02:14:19 crc kubenswrapper[4912]: E1203 02:14:19.274909 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="d5f2ea7f-cdc9-4573-8b16-6561c761f91f" Dec 03 02:14:28 crc kubenswrapper[4912]: I1203 02:14:28.572487 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:14:28 crc kubenswrapper[4912]: E1203 02:14:28.573773 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:14:33 crc kubenswrapper[4912]: I1203 02:14:33.446187 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d5f2ea7f-cdc9-4573-8b16-6561c761f91f","Type":"ContainerStarted","Data":"95ca4f945c84b30eb7280d9304ad12914e810485b59bd9ae9bdd4640819db579"} Dec 03 02:14:33 crc kubenswrapper[4912]: I1203 02:14:33.503948 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=5.084456201 podStartE2EDuration="56.503899949s" podCreationTimestamp="2025-12-03 02:13:37 +0000 UTC" firstStartedPulling="2025-12-03 02:13:39.83487636 +0000 UTC m=+6605.476896920" lastFinishedPulling="2025-12-03 02:14:31.254320068 +0000 UTC m=+6656.896340668" observedRunningTime="2025-12-03 02:14:33.479075233 +0000 UTC m=+6659.121095813" watchObservedRunningTime="2025-12-03 02:14:33.503899949 +0000 UTC m=+6659.145920579" Dec 03 02:14:40 crc kubenswrapper[4912]: I1203 02:14:40.572871 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:14:40 crc kubenswrapper[4912]: E1203 02:14:40.574075 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:14:55 crc kubenswrapper[4912]: I1203 02:14:55.572629 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:14:56 crc kubenswrapper[4912]: I1203 02:14:56.759021 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"6d754e8f4bdccaa6bcce5ba4114c2849fa8f32ea2ab9afbedf9e602eb587deb8"} Dec 03 02:14:58 crc kubenswrapper[4912]: I1203 02:14:58.847144 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-426kq"] Dec 03 02:14:58 crc kubenswrapper[4912]: I1203 02:14:58.850346 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:58 crc kubenswrapper[4912]: I1203 02:14:58.865931 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-426kq"] Dec 03 02:14:58 crc kubenswrapper[4912]: I1203 02:14:58.995567 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnltm\" (UniqueName: \"kubernetes.io/projected/52483bab-ddc1-4306-bbf9-4df03539c142-kube-api-access-bnltm\") pod \"redhat-marketplace-426kq\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:58 crc kubenswrapper[4912]: I1203 02:14:58.995982 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-utilities\") pod \"redhat-marketplace-426kq\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:58 crc kubenswrapper[4912]: I1203 02:14:58.996185 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-catalog-content\") pod \"redhat-marketplace-426kq\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:59 crc kubenswrapper[4912]: I1203 02:14:59.098280 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-catalog-content\") pod \"redhat-marketplace-426kq\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:59 crc kubenswrapper[4912]: I1203 02:14:59.098384 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnltm\" (UniqueName: \"kubernetes.io/projected/52483bab-ddc1-4306-bbf9-4df03539c142-kube-api-access-bnltm\") pod \"redhat-marketplace-426kq\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:59 crc kubenswrapper[4912]: I1203 02:14:59.098939 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-utilities\") pod \"redhat-marketplace-426kq\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:59 crc kubenswrapper[4912]: I1203 02:14:59.098947 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-catalog-content\") pod \"redhat-marketplace-426kq\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:59 crc kubenswrapper[4912]: I1203 02:14:59.099343 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-utilities\") pod \"redhat-marketplace-426kq\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:59 crc kubenswrapper[4912]: I1203 02:14:59.124655 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnltm\" (UniqueName: \"kubernetes.io/projected/52483bab-ddc1-4306-bbf9-4df03539c142-kube-api-access-bnltm\") pod \"redhat-marketplace-426kq\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:59 crc kubenswrapper[4912]: I1203 02:14:59.173402 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:14:59 crc kubenswrapper[4912]: I1203 02:14:59.738725 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-426kq"] Dec 03 02:14:59 crc kubenswrapper[4912]: I1203 02:14:59.809555 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-426kq" event={"ID":"52483bab-ddc1-4306-bbf9-4df03539c142","Type":"ContainerStarted","Data":"82d8b84de3e1eb206fb27c3b4a8a612807dea7d0e66611a99b88bd58319b3504"} Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.192363 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4"] Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.194000 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.199842 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.200011 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.202801 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4"] Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.344606 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3808d4a7-17e5-43c7-b253-21ac0607c155-config-volume\") pod \"collect-profiles-29412135-cn2b4\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.345476 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3808d4a7-17e5-43c7-b253-21ac0607c155-secret-volume\") pod \"collect-profiles-29412135-cn2b4\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.345762 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlhb8\" (UniqueName: \"kubernetes.io/projected/3808d4a7-17e5-43c7-b253-21ac0607c155-kube-api-access-vlhb8\") pod \"collect-profiles-29412135-cn2b4\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.448022 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3808d4a7-17e5-43c7-b253-21ac0607c155-config-volume\") pod \"collect-profiles-29412135-cn2b4\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.448117 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3808d4a7-17e5-43c7-b253-21ac0607c155-secret-volume\") pod \"collect-profiles-29412135-cn2b4\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.448232 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlhb8\" (UniqueName: \"kubernetes.io/projected/3808d4a7-17e5-43c7-b253-21ac0607c155-kube-api-access-vlhb8\") pod \"collect-profiles-29412135-cn2b4\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.448987 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3808d4a7-17e5-43c7-b253-21ac0607c155-config-volume\") pod \"collect-profiles-29412135-cn2b4\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.460171 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3808d4a7-17e5-43c7-b253-21ac0607c155-secret-volume\") pod \"collect-profiles-29412135-cn2b4\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.474270 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlhb8\" (UniqueName: \"kubernetes.io/projected/3808d4a7-17e5-43c7-b253-21ac0607c155-kube-api-access-vlhb8\") pod \"collect-profiles-29412135-cn2b4\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.525748 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.820492 4912 generic.go:334] "Generic (PLEG): container finished" podID="52483bab-ddc1-4306-bbf9-4df03539c142" containerID="8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5" exitCode=0 Dec 03 02:15:00 crc kubenswrapper[4912]: I1203 02:15:00.820579 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-426kq" event={"ID":"52483bab-ddc1-4306-bbf9-4df03539c142","Type":"ContainerDied","Data":"8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5"} Dec 03 02:15:01 crc kubenswrapper[4912]: I1203 02:15:01.082714 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4"] Dec 03 02:15:01 crc kubenswrapper[4912]: I1203 02:15:01.835760 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-426kq" event={"ID":"52483bab-ddc1-4306-bbf9-4df03539c142","Type":"ContainerStarted","Data":"232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe"} Dec 03 02:15:01 crc kubenswrapper[4912]: I1203 02:15:01.841773 4912 generic.go:334] "Generic (PLEG): container finished" podID="3808d4a7-17e5-43c7-b253-21ac0607c155" containerID="58f1ff9b2d182af77dff6c825e6b807560537299e0e33108914f1c9de37f7dc1" exitCode=0 Dec 03 02:15:01 crc kubenswrapper[4912]: I1203 02:15:01.841808 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" event={"ID":"3808d4a7-17e5-43c7-b253-21ac0607c155","Type":"ContainerDied","Data":"58f1ff9b2d182af77dff6c825e6b807560537299e0e33108914f1c9de37f7dc1"} Dec 03 02:15:01 crc kubenswrapper[4912]: I1203 02:15:01.841829 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" event={"ID":"3808d4a7-17e5-43c7-b253-21ac0607c155","Type":"ContainerStarted","Data":"adc40770475ab7e40ff3b9f53067ec226e1e2b36e2bcbeab7b05bda0ed2a8253"} Dec 03 02:15:02 crc kubenswrapper[4912]: I1203 02:15:02.865008 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-426kq" event={"ID":"52483bab-ddc1-4306-bbf9-4df03539c142","Type":"ContainerDied","Data":"232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe"} Dec 03 02:15:02 crc kubenswrapper[4912]: I1203 02:15:02.864844 4912 generic.go:334] "Generic (PLEG): container finished" podID="52483bab-ddc1-4306-bbf9-4df03539c142" containerID="232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe" exitCode=0 Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.375625 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.451557 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3808d4a7-17e5-43c7-b253-21ac0607c155-config-volume\") pod \"3808d4a7-17e5-43c7-b253-21ac0607c155\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.451631 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlhb8\" (UniqueName: \"kubernetes.io/projected/3808d4a7-17e5-43c7-b253-21ac0607c155-kube-api-access-vlhb8\") pod \"3808d4a7-17e5-43c7-b253-21ac0607c155\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.451886 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3808d4a7-17e5-43c7-b253-21ac0607c155-secret-volume\") pod \"3808d4a7-17e5-43c7-b253-21ac0607c155\" (UID: \"3808d4a7-17e5-43c7-b253-21ac0607c155\") " Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.452334 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3808d4a7-17e5-43c7-b253-21ac0607c155-config-volume" (OuterVolumeSpecName: "config-volume") pod "3808d4a7-17e5-43c7-b253-21ac0607c155" (UID: "3808d4a7-17e5-43c7-b253-21ac0607c155"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.452896 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3808d4a7-17e5-43c7-b253-21ac0607c155-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.457967 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3808d4a7-17e5-43c7-b253-21ac0607c155-kube-api-access-vlhb8" (OuterVolumeSpecName: "kube-api-access-vlhb8") pod "3808d4a7-17e5-43c7-b253-21ac0607c155" (UID: "3808d4a7-17e5-43c7-b253-21ac0607c155"). InnerVolumeSpecName "kube-api-access-vlhb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.458801 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3808d4a7-17e5-43c7-b253-21ac0607c155-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3808d4a7-17e5-43c7-b253-21ac0607c155" (UID: "3808d4a7-17e5-43c7-b253-21ac0607c155"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.555444 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3808d4a7-17e5-43c7-b253-21ac0607c155-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.555697 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlhb8\" (UniqueName: \"kubernetes.io/projected/3808d4a7-17e5-43c7-b253-21ac0607c155-kube-api-access-vlhb8\") on node \"crc\" DevicePath \"\"" Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.878792 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.878800 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412135-cn2b4" event={"ID":"3808d4a7-17e5-43c7-b253-21ac0607c155","Type":"ContainerDied","Data":"adc40770475ab7e40ff3b9f53067ec226e1e2b36e2bcbeab7b05bda0ed2a8253"} Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.880279 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adc40770475ab7e40ff3b9f53067ec226e1e2b36e2bcbeab7b05bda0ed2a8253" Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.882886 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-426kq" event={"ID":"52483bab-ddc1-4306-bbf9-4df03539c142","Type":"ContainerStarted","Data":"2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa"} Dec 03 02:15:03 crc kubenswrapper[4912]: I1203 02:15:03.915145 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-426kq" podStartSLOduration=3.396375922 podStartE2EDuration="5.915128897s" podCreationTimestamp="2025-12-03 02:14:58 +0000 UTC" firstStartedPulling="2025-12-03 02:15:00.822572963 +0000 UTC m=+6686.464593523" lastFinishedPulling="2025-12-03 02:15:03.341325938 +0000 UTC m=+6688.983346498" observedRunningTime="2025-12-03 02:15:03.910029791 +0000 UTC m=+6689.552050381" watchObservedRunningTime="2025-12-03 02:15:03.915128897 +0000 UTC m=+6689.557149457" Dec 03 02:15:04 crc kubenswrapper[4912]: I1203 02:15:04.459624 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w"] Dec 03 02:15:04 crc kubenswrapper[4912]: I1203 02:15:04.472728 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412090-hgj8w"] Dec 03 02:15:04 crc kubenswrapper[4912]: I1203 02:15:04.587953 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c01fab78-4545-4c05-a3c2-ed7d32cc222e" path="/var/lib/kubelet/pods/c01fab78-4545-4c05-a3c2-ed7d32cc222e/volumes" Dec 03 02:15:09 crc kubenswrapper[4912]: I1203 02:15:09.174166 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:15:09 crc kubenswrapper[4912]: I1203 02:15:09.174599 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:15:09 crc kubenswrapper[4912]: I1203 02:15:09.260220 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:15:10 crc kubenswrapper[4912]: I1203 02:15:10.039379 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:15:10 crc kubenswrapper[4912]: I1203 02:15:10.091489 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-426kq"] Dec 03 02:15:11 crc kubenswrapper[4912]: I1203 02:15:11.992999 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-426kq" podUID="52483bab-ddc1-4306-bbf9-4df03539c142" containerName="registry-server" containerID="cri-o://2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa" gracePeriod=2 Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.626592 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.799907 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnltm\" (UniqueName: \"kubernetes.io/projected/52483bab-ddc1-4306-bbf9-4df03539c142-kube-api-access-bnltm\") pod \"52483bab-ddc1-4306-bbf9-4df03539c142\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.800174 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-catalog-content\") pod \"52483bab-ddc1-4306-bbf9-4df03539c142\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.800320 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-utilities\") pod \"52483bab-ddc1-4306-bbf9-4df03539c142\" (UID: \"52483bab-ddc1-4306-bbf9-4df03539c142\") " Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.801835 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-utilities" (OuterVolumeSpecName: "utilities") pod "52483bab-ddc1-4306-bbf9-4df03539c142" (UID: "52483bab-ddc1-4306-bbf9-4df03539c142"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.819930 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52483bab-ddc1-4306-bbf9-4df03539c142-kube-api-access-bnltm" (OuterVolumeSpecName: "kube-api-access-bnltm") pod "52483bab-ddc1-4306-bbf9-4df03539c142" (UID: "52483bab-ddc1-4306-bbf9-4df03539c142"). InnerVolumeSpecName "kube-api-access-bnltm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.827106 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52483bab-ddc1-4306-bbf9-4df03539c142" (UID: "52483bab-ddc1-4306-bbf9-4df03539c142"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.903103 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnltm\" (UniqueName: \"kubernetes.io/projected/52483bab-ddc1-4306-bbf9-4df03539c142-kube-api-access-bnltm\") on node \"crc\" DevicePath \"\"" Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.903375 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:15:12 crc kubenswrapper[4912]: I1203 02:15:12.903385 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52483bab-ddc1-4306-bbf9-4df03539c142-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.001414 4912 generic.go:334] "Generic (PLEG): container finished" podID="52483bab-ddc1-4306-bbf9-4df03539c142" containerID="2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa" exitCode=0 Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.001463 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-426kq" event={"ID":"52483bab-ddc1-4306-bbf9-4df03539c142","Type":"ContainerDied","Data":"2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa"} Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.001487 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-426kq" event={"ID":"52483bab-ddc1-4306-bbf9-4df03539c142","Type":"ContainerDied","Data":"82d8b84de3e1eb206fb27c3b4a8a612807dea7d0e66611a99b88bd58319b3504"} Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.001516 4912 scope.go:117] "RemoveContainer" containerID="2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.001643 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-426kq" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.035630 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-426kq"] Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.044577 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-426kq"] Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.046706 4912 scope.go:117] "RemoveContainer" containerID="232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.074647 4912 scope.go:117] "RemoveContainer" containerID="8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.144364 4912 scope.go:117] "RemoveContainer" containerID="2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa" Dec 03 02:15:13 crc kubenswrapper[4912]: E1203 02:15:13.144909 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa\": container with ID starting with 2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa not found: ID does not exist" containerID="2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.144950 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa"} err="failed to get container status \"2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa\": rpc error: code = NotFound desc = could not find container \"2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa\": container with ID starting with 2a106b085c8cbc830bed69b3d627f7640ee6fcaebf7d4b2ae5cd814d9fe890aa not found: ID does not exist" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.144994 4912 scope.go:117] "RemoveContainer" containerID="232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe" Dec 03 02:15:13 crc kubenswrapper[4912]: E1203 02:15:13.145289 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe\": container with ID starting with 232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe not found: ID does not exist" containerID="232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.145357 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe"} err="failed to get container status \"232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe\": rpc error: code = NotFound desc = could not find container \"232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe\": container with ID starting with 232351d309b962498aa7a6da7605ce96f1faeb85db90ff81435847b3024defbe not found: ID does not exist" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.145385 4912 scope.go:117] "RemoveContainer" containerID="8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5" Dec 03 02:15:13 crc kubenswrapper[4912]: E1203 02:15:13.145893 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5\": container with ID starting with 8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5 not found: ID does not exist" containerID="8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5" Dec 03 02:15:13 crc kubenswrapper[4912]: I1203 02:15:13.145931 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5"} err="failed to get container status \"8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5\": rpc error: code = NotFound desc = could not find container \"8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5\": container with ID starting with 8b74f5edfab8852bc554f2f1476d561967845c41c28e5412d2e16748a70416e5 not found: ID does not exist" Dec 03 02:15:14 crc kubenswrapper[4912]: I1203 02:15:14.587564 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52483bab-ddc1-4306-bbf9-4df03539c142" path="/var/lib/kubelet/pods/52483bab-ddc1-4306-bbf9-4df03539c142/volumes" Dec 03 02:15:28 crc kubenswrapper[4912]: I1203 02:15:28.636176 4912 scope.go:117] "RemoveContainer" containerID="969f718bec11c2f1e6ad4f4790843029de5ae39182502f83370ee93ed37f4cf8" Dec 03 02:16:56 crc kubenswrapper[4912]: E1203 02:16:56.080767 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Dec 03 02:17:18 crc kubenswrapper[4912]: I1203 02:17:18.078590 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:17:18 crc kubenswrapper[4912]: I1203 02:17:18.080533 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:17:48 crc kubenswrapper[4912]: I1203 02:17:48.077994 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:17:48 crc kubenswrapper[4912]: I1203 02:17:48.078529 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:18:18 crc kubenswrapper[4912]: I1203 02:18:18.077964 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:18:18 crc kubenswrapper[4912]: I1203 02:18:18.078657 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:18:18 crc kubenswrapper[4912]: I1203 02:18:18.078721 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:18:18 crc kubenswrapper[4912]: I1203 02:18:18.080355 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6d754e8f4bdccaa6bcce5ba4114c2849fa8f32ea2ab9afbedf9e602eb587deb8"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:18:18 crc kubenswrapper[4912]: I1203 02:18:18.080470 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://6d754e8f4bdccaa6bcce5ba4114c2849fa8f32ea2ab9afbedf9e602eb587deb8" gracePeriod=600 Dec 03 02:18:19 crc kubenswrapper[4912]: I1203 02:18:19.221730 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="6d754e8f4bdccaa6bcce5ba4114c2849fa8f32ea2ab9afbedf9e602eb587deb8" exitCode=0 Dec 03 02:18:19 crc kubenswrapper[4912]: I1203 02:18:19.221807 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"6d754e8f4bdccaa6bcce5ba4114c2849fa8f32ea2ab9afbedf9e602eb587deb8"} Dec 03 02:18:19 crc kubenswrapper[4912]: I1203 02:18:19.222160 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5"} Dec 03 02:18:19 crc kubenswrapper[4912]: I1203 02:18:19.222178 4912 scope.go:117] "RemoveContainer" containerID="dbb0265fc48e0d85e99a0dee703b50211410797821c175ba126988724feef2da" Dec 03 02:19:17 crc kubenswrapper[4912]: I1203 02:19:17.926959 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mm9fc"] Dec 03 02:19:17 crc kubenswrapper[4912]: E1203 02:19:17.928107 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52483bab-ddc1-4306-bbf9-4df03539c142" containerName="registry-server" Dec 03 02:19:17 crc kubenswrapper[4912]: I1203 02:19:17.928133 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="52483bab-ddc1-4306-bbf9-4df03539c142" containerName="registry-server" Dec 03 02:19:17 crc kubenswrapper[4912]: E1203 02:19:17.928152 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3808d4a7-17e5-43c7-b253-21ac0607c155" containerName="collect-profiles" Dec 03 02:19:17 crc kubenswrapper[4912]: I1203 02:19:17.928158 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="3808d4a7-17e5-43c7-b253-21ac0607c155" containerName="collect-profiles" Dec 03 02:19:17 crc kubenswrapper[4912]: E1203 02:19:17.928168 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52483bab-ddc1-4306-bbf9-4df03539c142" containerName="extract-content" Dec 03 02:19:17 crc kubenswrapper[4912]: I1203 02:19:17.928175 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="52483bab-ddc1-4306-bbf9-4df03539c142" containerName="extract-content" Dec 03 02:19:17 crc kubenswrapper[4912]: E1203 02:19:17.928197 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52483bab-ddc1-4306-bbf9-4df03539c142" containerName="extract-utilities" Dec 03 02:19:17 crc kubenswrapper[4912]: I1203 02:19:17.928203 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="52483bab-ddc1-4306-bbf9-4df03539c142" containerName="extract-utilities" Dec 03 02:19:17 crc kubenswrapper[4912]: I1203 02:19:17.928909 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="52483bab-ddc1-4306-bbf9-4df03539c142" containerName="registry-server" Dec 03 02:19:17 crc kubenswrapper[4912]: I1203 02:19:17.928926 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="3808d4a7-17e5-43c7-b253-21ac0607c155" containerName="collect-profiles" Dec 03 02:19:17 crc kubenswrapper[4912]: I1203 02:19:17.930686 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:17 crc kubenswrapper[4912]: I1203 02:19:17.947830 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mm9fc"] Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.000833 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-catalog-content\") pod \"certified-operators-mm9fc\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.001089 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w68x7\" (UniqueName: \"kubernetes.io/projected/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-kube-api-access-w68x7\") pod \"certified-operators-mm9fc\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.001239 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-utilities\") pod \"certified-operators-mm9fc\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.103736 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w68x7\" (UniqueName: \"kubernetes.io/projected/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-kube-api-access-w68x7\") pod \"certified-operators-mm9fc\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.103850 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-utilities\") pod \"certified-operators-mm9fc\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.103888 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-catalog-content\") pod \"certified-operators-mm9fc\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.104622 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-catalog-content\") pod \"certified-operators-mm9fc\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.104830 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-utilities\") pod \"certified-operators-mm9fc\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.124156 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w68x7\" (UniqueName: \"kubernetes.io/projected/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-kube-api-access-w68x7\") pod \"certified-operators-mm9fc\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.255597 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:18 crc kubenswrapper[4912]: I1203 02:19:18.848903 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mm9fc"] Dec 03 02:19:19 crc kubenswrapper[4912]: I1203 02:19:19.181121 4912 generic.go:334] "Generic (PLEG): container finished" podID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerID="a504b46e0ebd871c0ee6b5f8c9ac14c9af90acb75764d2b6f3aea60a5a4b30e1" exitCode=0 Dec 03 02:19:19 crc kubenswrapper[4912]: I1203 02:19:19.181169 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm9fc" event={"ID":"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82","Type":"ContainerDied","Data":"a504b46e0ebd871c0ee6b5f8c9ac14c9af90acb75764d2b6f3aea60a5a4b30e1"} Dec 03 02:19:19 crc kubenswrapper[4912]: I1203 02:19:19.181384 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm9fc" event={"ID":"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82","Type":"ContainerStarted","Data":"930b79c08561fd5385fc6e2c63bf7449a2b66dfe9590d69a9183cf3d61f82fd4"} Dec 03 02:19:19 crc kubenswrapper[4912]: I1203 02:19:19.184685 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 02:19:20 crc kubenswrapper[4912]: I1203 02:19:20.191794 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm9fc" event={"ID":"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82","Type":"ContainerStarted","Data":"5ec326941d1ce1c247429081aa97fd074ed86a3becc42fe00c262f21a4bc2519"} Dec 03 02:19:21 crc kubenswrapper[4912]: I1203 02:19:21.202914 4912 generic.go:334] "Generic (PLEG): container finished" podID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerID="5ec326941d1ce1c247429081aa97fd074ed86a3becc42fe00c262f21a4bc2519" exitCode=0 Dec 03 02:19:21 crc kubenswrapper[4912]: I1203 02:19:21.202966 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm9fc" event={"ID":"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82","Type":"ContainerDied","Data":"5ec326941d1ce1c247429081aa97fd074ed86a3becc42fe00c262f21a4bc2519"} Dec 03 02:19:22 crc kubenswrapper[4912]: I1203 02:19:22.216646 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm9fc" event={"ID":"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82","Type":"ContainerStarted","Data":"c37c81be241ef18743f44636e27c2327f99610bd41f3e42bd1f4c64feb290aa0"} Dec 03 02:19:22 crc kubenswrapper[4912]: I1203 02:19:22.236395 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mm9fc" podStartSLOduration=2.61703699 podStartE2EDuration="5.236379365s" podCreationTimestamp="2025-12-03 02:19:17 +0000 UTC" firstStartedPulling="2025-12-03 02:19:19.184295417 +0000 UTC m=+6944.826315977" lastFinishedPulling="2025-12-03 02:19:21.803637792 +0000 UTC m=+6947.445658352" observedRunningTime="2025-12-03 02:19:22.233372245 +0000 UTC m=+6947.875392805" watchObservedRunningTime="2025-12-03 02:19:22.236379365 +0000 UTC m=+6947.878399925" Dec 03 02:19:28 crc kubenswrapper[4912]: I1203 02:19:28.256442 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:28 crc kubenswrapper[4912]: I1203 02:19:28.256843 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:28 crc kubenswrapper[4912]: I1203 02:19:28.310951 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:28 crc kubenswrapper[4912]: I1203 02:19:28.373172 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:28 crc kubenswrapper[4912]: I1203 02:19:28.548044 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mm9fc"] Dec 03 02:19:30 crc kubenswrapper[4912]: I1203 02:19:30.308364 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mm9fc" podUID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerName="registry-server" containerID="cri-o://c37c81be241ef18743f44636e27c2327f99610bd41f3e42bd1f4c64feb290aa0" gracePeriod=2 Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.322615 4912 generic.go:334] "Generic (PLEG): container finished" podID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerID="c37c81be241ef18743f44636e27c2327f99610bd41f3e42bd1f4c64feb290aa0" exitCode=0 Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.322751 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm9fc" event={"ID":"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82","Type":"ContainerDied","Data":"c37c81be241ef18743f44636e27c2327f99610bd41f3e42bd1f4c64feb290aa0"} Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.484143 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.671678 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-utilities\") pod \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.671805 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w68x7\" (UniqueName: \"kubernetes.io/projected/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-kube-api-access-w68x7\") pod \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.672004 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-catalog-content\") pod \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\" (UID: \"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82\") " Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.672548 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-utilities" (OuterVolumeSpecName: "utilities") pod "8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" (UID: "8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.681050 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-kube-api-access-w68x7" (OuterVolumeSpecName: "kube-api-access-w68x7") pod "8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" (UID: "8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82"). InnerVolumeSpecName "kube-api-access-w68x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.717494 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" (UID: "8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.775207 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w68x7\" (UniqueName: \"kubernetes.io/projected/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-kube-api-access-w68x7\") on node \"crc\" DevicePath \"\"" Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.775249 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:19:31 crc kubenswrapper[4912]: I1203 02:19:31.775264 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:19:32 crc kubenswrapper[4912]: I1203 02:19:32.342470 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mm9fc" event={"ID":"8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82","Type":"ContainerDied","Data":"930b79c08561fd5385fc6e2c63bf7449a2b66dfe9590d69a9183cf3d61f82fd4"} Dec 03 02:19:32 crc kubenswrapper[4912]: I1203 02:19:32.342580 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mm9fc" Dec 03 02:19:32 crc kubenswrapper[4912]: I1203 02:19:32.342902 4912 scope.go:117] "RemoveContainer" containerID="c37c81be241ef18743f44636e27c2327f99610bd41f3e42bd1f4c64feb290aa0" Dec 03 02:19:32 crc kubenswrapper[4912]: I1203 02:19:32.378940 4912 scope.go:117] "RemoveContainer" containerID="5ec326941d1ce1c247429081aa97fd074ed86a3becc42fe00c262f21a4bc2519" Dec 03 02:19:32 crc kubenswrapper[4912]: I1203 02:19:32.395735 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mm9fc"] Dec 03 02:19:32 crc kubenswrapper[4912]: I1203 02:19:32.417850 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mm9fc"] Dec 03 02:19:32 crc kubenswrapper[4912]: I1203 02:19:32.427245 4912 scope.go:117] "RemoveContainer" containerID="a504b46e0ebd871c0ee6b5f8c9ac14c9af90acb75764d2b6f3aea60a5a4b30e1" Dec 03 02:19:32 crc kubenswrapper[4912]: I1203 02:19:32.591456 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" path="/var/lib/kubelet/pods/8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82/volumes" Dec 03 02:20:18 crc kubenswrapper[4912]: I1203 02:20:18.077919 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:20:18 crc kubenswrapper[4912]: I1203 02:20:18.078601 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.385116 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mglg5"] Dec 03 02:20:32 crc kubenswrapper[4912]: E1203 02:20:32.386063 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerName="extract-utilities" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.386077 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerName="extract-utilities" Dec 03 02:20:32 crc kubenswrapper[4912]: E1203 02:20:32.386104 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerName="extract-content" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.386110 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerName="extract-content" Dec 03 02:20:32 crc kubenswrapper[4912]: E1203 02:20:32.386137 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerName="registry-server" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.386145 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerName="registry-server" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.386373 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bcb7f9d-76ec-4d64-8c86-7ba0446d6e82" containerName="registry-server" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.387992 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.416686 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mglg5"] Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.491779 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-utilities\") pod \"community-operators-mglg5\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.491916 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9t9j\" (UniqueName: \"kubernetes.io/projected/3266b492-f2dc-42a4-86a5-1d52438532b8-kube-api-access-q9t9j\") pod \"community-operators-mglg5\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.492241 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-catalog-content\") pod \"community-operators-mglg5\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.593386 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-catalog-content\") pod \"community-operators-mglg5\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.594152 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-utilities\") pod \"community-operators-mglg5\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.594157 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-catalog-content\") pod \"community-operators-mglg5\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.594209 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9t9j\" (UniqueName: \"kubernetes.io/projected/3266b492-f2dc-42a4-86a5-1d52438532b8-kube-api-access-q9t9j\") pod \"community-operators-mglg5\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.594362 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-utilities\") pod \"community-operators-mglg5\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.621191 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9t9j\" (UniqueName: \"kubernetes.io/projected/3266b492-f2dc-42a4-86a5-1d52438532b8-kube-api-access-q9t9j\") pod \"community-operators-mglg5\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:32 crc kubenswrapper[4912]: I1203 02:20:32.769678 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.323398 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mglg5"] Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.398261 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hwfkr"] Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.426765 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.427642 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwfkr"] Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.625379 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-utilities\") pod \"redhat-operators-hwfkr\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.625740 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnqxd\" (UniqueName: \"kubernetes.io/projected/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-kube-api-access-dnqxd\") pod \"redhat-operators-hwfkr\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.625832 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-catalog-content\") pod \"redhat-operators-hwfkr\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.727811 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-catalog-content\") pod \"redhat-operators-hwfkr\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.727962 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-utilities\") pod \"redhat-operators-hwfkr\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.728009 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnqxd\" (UniqueName: \"kubernetes.io/projected/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-kube-api-access-dnqxd\") pod \"redhat-operators-hwfkr\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.728446 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-utilities\") pod \"redhat-operators-hwfkr\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.728451 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-catalog-content\") pod \"redhat-operators-hwfkr\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.750456 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnqxd\" (UniqueName: \"kubernetes.io/projected/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-kube-api-access-dnqxd\") pod \"redhat-operators-hwfkr\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:33 crc kubenswrapper[4912]: I1203 02:20:33.785253 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:34 crc kubenswrapper[4912]: I1203 02:20:34.118889 4912 generic.go:334] "Generic (PLEG): container finished" podID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerID="5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c" exitCode=0 Dec 03 02:20:34 crc kubenswrapper[4912]: I1203 02:20:34.119029 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mglg5" event={"ID":"3266b492-f2dc-42a4-86a5-1d52438532b8","Type":"ContainerDied","Data":"5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c"} Dec 03 02:20:34 crc kubenswrapper[4912]: I1203 02:20:34.119140 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mglg5" event={"ID":"3266b492-f2dc-42a4-86a5-1d52438532b8","Type":"ContainerStarted","Data":"266ad5d86b76653bef14bb0658181b0edcbffe8a9b5e402bd3ab87b98e836f85"} Dec 03 02:20:34 crc kubenswrapper[4912]: I1203 02:20:34.264544 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hwfkr"] Dec 03 02:20:35 crc kubenswrapper[4912]: I1203 02:20:35.130274 4912 generic.go:334] "Generic (PLEG): container finished" podID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerID="ca1a951da922d82123a08e4864070bacf5fe46ed9269a172f400bab989996ab5" exitCode=0 Dec 03 02:20:35 crc kubenswrapper[4912]: I1203 02:20:35.130501 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwfkr" event={"ID":"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5","Type":"ContainerDied","Data":"ca1a951da922d82123a08e4864070bacf5fe46ed9269a172f400bab989996ab5"} Dec 03 02:20:35 crc kubenswrapper[4912]: I1203 02:20:35.130850 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwfkr" event={"ID":"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5","Type":"ContainerStarted","Data":"da39d61842fa3afe739e445472faa03841df34f479aa3603f61c15be010b98f3"} Dec 03 02:20:35 crc kubenswrapper[4912]: I1203 02:20:35.133423 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mglg5" event={"ID":"3266b492-f2dc-42a4-86a5-1d52438532b8","Type":"ContainerStarted","Data":"2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6"} Dec 03 02:20:37 crc kubenswrapper[4912]: I1203 02:20:37.160749 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwfkr" event={"ID":"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5","Type":"ContainerStarted","Data":"b9cb2784d585ac924b0be2e909333e9511cb030cbb2e71b26bb74dfec5bb2425"} Dec 03 02:20:37 crc kubenswrapper[4912]: I1203 02:20:37.166989 4912 generic.go:334] "Generic (PLEG): container finished" podID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerID="2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6" exitCode=0 Dec 03 02:20:37 crc kubenswrapper[4912]: I1203 02:20:37.167171 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mglg5" event={"ID":"3266b492-f2dc-42a4-86a5-1d52438532b8","Type":"ContainerDied","Data":"2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6"} Dec 03 02:20:38 crc kubenswrapper[4912]: I1203 02:20:38.181316 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mglg5" event={"ID":"3266b492-f2dc-42a4-86a5-1d52438532b8","Type":"ContainerStarted","Data":"84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4"} Dec 03 02:20:38 crc kubenswrapper[4912]: I1203 02:20:38.204333 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mglg5" podStartSLOduration=2.613351238 podStartE2EDuration="6.204314353s" podCreationTimestamp="2025-12-03 02:20:32 +0000 UTC" firstStartedPulling="2025-12-03 02:20:34.122766858 +0000 UTC m=+7019.764787418" lastFinishedPulling="2025-12-03 02:20:37.713729973 +0000 UTC m=+7023.355750533" observedRunningTime="2025-12-03 02:20:38.196496412 +0000 UTC m=+7023.838516982" watchObservedRunningTime="2025-12-03 02:20:38.204314353 +0000 UTC m=+7023.846334913" Dec 03 02:20:39 crc kubenswrapper[4912]: I1203 02:20:39.193024 4912 generic.go:334] "Generic (PLEG): container finished" podID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerID="b9cb2784d585ac924b0be2e909333e9511cb030cbb2e71b26bb74dfec5bb2425" exitCode=0 Dec 03 02:20:39 crc kubenswrapper[4912]: I1203 02:20:39.193079 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwfkr" event={"ID":"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5","Type":"ContainerDied","Data":"b9cb2784d585ac924b0be2e909333e9511cb030cbb2e71b26bb74dfec5bb2425"} Dec 03 02:20:40 crc kubenswrapper[4912]: I1203 02:20:40.205231 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwfkr" event={"ID":"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5","Type":"ContainerStarted","Data":"262f18c72b01113df2a9f66937e84166883c00c46e74a69ec1836320ba304925"} Dec 03 02:20:40 crc kubenswrapper[4912]: I1203 02:20:40.233409 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hwfkr" podStartSLOduration=2.734730257 podStartE2EDuration="7.23338703s" podCreationTimestamp="2025-12-03 02:20:33 +0000 UTC" firstStartedPulling="2025-12-03 02:20:35.131833447 +0000 UTC m=+7020.773854007" lastFinishedPulling="2025-12-03 02:20:39.63049022 +0000 UTC m=+7025.272510780" observedRunningTime="2025-12-03 02:20:40.223573541 +0000 UTC m=+7025.865594111" watchObservedRunningTime="2025-12-03 02:20:40.23338703 +0000 UTC m=+7025.875407590" Dec 03 02:20:42 crc kubenswrapper[4912]: I1203 02:20:42.772600 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:42 crc kubenswrapper[4912]: I1203 02:20:42.773151 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:43 crc kubenswrapper[4912]: I1203 02:20:43.785934 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:43 crc kubenswrapper[4912]: I1203 02:20:43.786328 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:43 crc kubenswrapper[4912]: I1203 02:20:43.840010 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mglg5" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerName="registry-server" probeResult="failure" output=< Dec 03 02:20:43 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 02:20:43 crc kubenswrapper[4912]: > Dec 03 02:20:44 crc kubenswrapper[4912]: I1203 02:20:44.854053 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hwfkr" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerName="registry-server" probeResult="failure" output=< Dec 03 02:20:44 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 02:20:44 crc kubenswrapper[4912]: > Dec 03 02:20:48 crc kubenswrapper[4912]: I1203 02:20:48.078028 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:20:48 crc kubenswrapper[4912]: I1203 02:20:48.078930 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:20:52 crc kubenswrapper[4912]: I1203 02:20:52.909106 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:52 crc kubenswrapper[4912]: I1203 02:20:52.999713 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:53 crc kubenswrapper[4912]: I1203 02:20:53.156771 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mglg5"] Dec 03 02:20:53 crc kubenswrapper[4912]: I1203 02:20:53.881053 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:53 crc kubenswrapper[4912]: I1203 02:20:53.956397 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:54 crc kubenswrapper[4912]: I1203 02:20:54.363298 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mglg5" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerName="registry-server" containerID="cri-o://84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4" gracePeriod=2 Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.027538 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.133669 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-catalog-content\") pod \"3266b492-f2dc-42a4-86a5-1d52438532b8\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.133804 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9t9j\" (UniqueName: \"kubernetes.io/projected/3266b492-f2dc-42a4-86a5-1d52438532b8-kube-api-access-q9t9j\") pod \"3266b492-f2dc-42a4-86a5-1d52438532b8\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.134057 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-utilities\") pod \"3266b492-f2dc-42a4-86a5-1d52438532b8\" (UID: \"3266b492-f2dc-42a4-86a5-1d52438532b8\") " Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.135061 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-utilities" (OuterVolumeSpecName: "utilities") pod "3266b492-f2dc-42a4-86a5-1d52438532b8" (UID: "3266b492-f2dc-42a4-86a5-1d52438532b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.154757 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3266b492-f2dc-42a4-86a5-1d52438532b8-kube-api-access-q9t9j" (OuterVolumeSpecName: "kube-api-access-q9t9j") pod "3266b492-f2dc-42a4-86a5-1d52438532b8" (UID: "3266b492-f2dc-42a4-86a5-1d52438532b8"). InnerVolumeSpecName "kube-api-access-q9t9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.195616 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3266b492-f2dc-42a4-86a5-1d52438532b8" (UID: "3266b492-f2dc-42a4-86a5-1d52438532b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.237304 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.237354 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3266b492-f2dc-42a4-86a5-1d52438532b8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.237660 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9t9j\" (UniqueName: \"kubernetes.io/projected/3266b492-f2dc-42a4-86a5-1d52438532b8-kube-api-access-q9t9j\") on node \"crc\" DevicePath \"\"" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.376108 4912 generic.go:334] "Generic (PLEG): container finished" podID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerID="84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4" exitCode=0 Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.376208 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mglg5" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.376234 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mglg5" event={"ID":"3266b492-f2dc-42a4-86a5-1d52438532b8","Type":"ContainerDied","Data":"84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4"} Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.376525 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mglg5" event={"ID":"3266b492-f2dc-42a4-86a5-1d52438532b8","Type":"ContainerDied","Data":"266ad5d86b76653bef14bb0658181b0edcbffe8a9b5e402bd3ab87b98e836f85"} Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.376565 4912 scope.go:117] "RemoveContainer" containerID="84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.412206 4912 scope.go:117] "RemoveContainer" containerID="2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.423228 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mglg5"] Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.434891 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mglg5"] Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.453816 4912 scope.go:117] "RemoveContainer" containerID="5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.510616 4912 scope.go:117] "RemoveContainer" containerID="84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4" Dec 03 02:20:55 crc kubenswrapper[4912]: E1203 02:20:55.513906 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4\": container with ID starting with 84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4 not found: ID does not exist" containerID="84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.513952 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4"} err="failed to get container status \"84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4\": rpc error: code = NotFound desc = could not find container \"84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4\": container with ID starting with 84cfd00e5a5cd8119fa965a61941b2fba947b4be6080421012d05244d17bb0e4 not found: ID does not exist" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.513997 4912 scope.go:117] "RemoveContainer" containerID="2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6" Dec 03 02:20:55 crc kubenswrapper[4912]: E1203 02:20:55.514476 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6\": container with ID starting with 2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6 not found: ID does not exist" containerID="2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.514519 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6"} err="failed to get container status \"2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6\": rpc error: code = NotFound desc = could not find container \"2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6\": container with ID starting with 2777677e44ba553cb30a0cc90b683e6184ad5e6770532251543a4091afee52d6 not found: ID does not exist" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.514546 4912 scope.go:117] "RemoveContainer" containerID="5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c" Dec 03 02:20:55 crc kubenswrapper[4912]: E1203 02:20:55.514841 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c\": container with ID starting with 5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c not found: ID does not exist" containerID="5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.514877 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c"} err="failed to get container status \"5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c\": rpc error: code = NotFound desc = could not find container \"5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c\": container with ID starting with 5bd12e28d0070dee3ed2def221031449bf6be0a877d0ad3ec636b2a3cf2ecb1c not found: ID does not exist" Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.966735 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hwfkr"] Dec 03 02:20:55 crc kubenswrapper[4912]: I1203 02:20:55.967751 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hwfkr" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerName="registry-server" containerID="cri-o://262f18c72b01113df2a9f66937e84166883c00c46e74a69ec1836320ba304925" gracePeriod=2 Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.388289 4912 generic.go:334] "Generic (PLEG): container finished" podID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerID="262f18c72b01113df2a9f66937e84166883c00c46e74a69ec1836320ba304925" exitCode=0 Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.388542 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwfkr" event={"ID":"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5","Type":"ContainerDied","Data":"262f18c72b01113df2a9f66937e84166883c00c46e74a69ec1836320ba304925"} Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.556641 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.609366 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" path="/var/lib/kubelet/pods/3266b492-f2dc-42a4-86a5-1d52438532b8/volumes" Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.676127 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-catalog-content\") pod \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.676349 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-utilities\") pod \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.676387 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnqxd\" (UniqueName: \"kubernetes.io/projected/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-kube-api-access-dnqxd\") pod \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\" (UID: \"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5\") " Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.680325 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-utilities" (OuterVolumeSpecName: "utilities") pod "dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" (UID: "dc7b7639-cebd-4b44-b0fe-5cffdcc227d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.688530 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-kube-api-access-dnqxd" (OuterVolumeSpecName: "kube-api-access-dnqxd") pod "dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" (UID: "dc7b7639-cebd-4b44-b0fe-5cffdcc227d5"). InnerVolumeSpecName "kube-api-access-dnqxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.779166 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.779204 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnqxd\" (UniqueName: \"kubernetes.io/projected/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-kube-api-access-dnqxd\") on node \"crc\" DevicePath \"\"" Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.814982 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" (UID: "dc7b7639-cebd-4b44-b0fe-5cffdcc227d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:20:56 crc kubenswrapper[4912]: I1203 02:20:56.881358 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:20:57 crc kubenswrapper[4912]: I1203 02:20:57.405459 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hwfkr" event={"ID":"dc7b7639-cebd-4b44-b0fe-5cffdcc227d5","Type":"ContainerDied","Data":"da39d61842fa3afe739e445472faa03841df34f479aa3603f61c15be010b98f3"} Dec 03 02:20:57 crc kubenswrapper[4912]: I1203 02:20:57.405511 4912 scope.go:117] "RemoveContainer" containerID="262f18c72b01113df2a9f66937e84166883c00c46e74a69ec1836320ba304925" Dec 03 02:20:57 crc kubenswrapper[4912]: I1203 02:20:57.405513 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hwfkr" Dec 03 02:20:57 crc kubenswrapper[4912]: I1203 02:20:57.439287 4912 scope.go:117] "RemoveContainer" containerID="b9cb2784d585ac924b0be2e909333e9511cb030cbb2e71b26bb74dfec5bb2425" Dec 03 02:20:57 crc kubenswrapper[4912]: I1203 02:20:57.453775 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hwfkr"] Dec 03 02:20:57 crc kubenswrapper[4912]: I1203 02:20:57.468578 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hwfkr"] Dec 03 02:20:57 crc kubenswrapper[4912]: I1203 02:20:57.470258 4912 scope.go:117] "RemoveContainer" containerID="ca1a951da922d82123a08e4864070bacf5fe46ed9269a172f400bab989996ab5" Dec 03 02:20:58 crc kubenswrapper[4912]: I1203 02:20:58.612752 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" path="/var/lib/kubelet/pods/dc7b7639-cebd-4b44-b0fe-5cffdcc227d5/volumes" Dec 03 02:21:18 crc kubenswrapper[4912]: I1203 02:21:18.078518 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:21:18 crc kubenswrapper[4912]: I1203 02:21:18.079271 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:21:18 crc kubenswrapper[4912]: I1203 02:21:18.079346 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:21:18 crc kubenswrapper[4912]: I1203 02:21:18.080719 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:21:18 crc kubenswrapper[4912]: I1203 02:21:18.080834 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" gracePeriod=600 Dec 03 02:21:18 crc kubenswrapper[4912]: E1203 02:21:18.208507 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:21:18 crc kubenswrapper[4912]: I1203 02:21:18.699701 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" exitCode=0 Dec 03 02:21:18 crc kubenswrapper[4912]: I1203 02:21:18.699773 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5"} Dec 03 02:21:18 crc kubenswrapper[4912]: I1203 02:21:18.699832 4912 scope.go:117] "RemoveContainer" containerID="6d754e8f4bdccaa6bcce5ba4114c2849fa8f32ea2ab9afbedf9e602eb587deb8" Dec 03 02:21:18 crc kubenswrapper[4912]: I1203 02:21:18.701014 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:21:18 crc kubenswrapper[4912]: E1203 02:21:18.701756 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:21:32 crc kubenswrapper[4912]: I1203 02:21:32.571931 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:21:32 crc kubenswrapper[4912]: E1203 02:21:32.572866 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:21:45 crc kubenswrapper[4912]: I1203 02:21:45.572166 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:21:45 crc kubenswrapper[4912]: E1203 02:21:45.573003 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:21:59 crc kubenswrapper[4912]: I1203 02:21:59.571841 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:21:59 crc kubenswrapper[4912]: E1203 02:21:59.572761 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:22:12 crc kubenswrapper[4912]: I1203 02:22:12.571931 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:22:12 crc kubenswrapper[4912]: E1203 02:22:12.572693 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:22:24 crc kubenswrapper[4912]: I1203 02:22:24.589206 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:22:24 crc kubenswrapper[4912]: E1203 02:22:24.590397 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:22:36 crc kubenswrapper[4912]: I1203 02:22:36.572135 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:22:36 crc kubenswrapper[4912]: E1203 02:22:36.573063 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:22:50 crc kubenswrapper[4912]: I1203 02:22:50.571480 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:22:50 crc kubenswrapper[4912]: E1203 02:22:50.572336 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:23:05 crc kubenswrapper[4912]: I1203 02:23:05.572793 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:23:05 crc kubenswrapper[4912]: E1203 02:23:05.573931 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:23:17 crc kubenswrapper[4912]: I1203 02:23:17.572599 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:23:17 crc kubenswrapper[4912]: E1203 02:23:17.573498 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:23:31 crc kubenswrapper[4912]: I1203 02:23:31.572661 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:23:31 crc kubenswrapper[4912]: E1203 02:23:31.573933 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:23:44 crc kubenswrapper[4912]: I1203 02:23:44.586927 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:23:44 crc kubenswrapper[4912]: E1203 02:23:44.587835 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:23:55 crc kubenswrapper[4912]: I1203 02:23:55.571524 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:23:55 crc kubenswrapper[4912]: E1203 02:23:55.572303 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:24:06 crc kubenswrapper[4912]: I1203 02:24:06.572337 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:24:06 crc kubenswrapper[4912]: E1203 02:24:06.574367 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:24:18 crc kubenswrapper[4912]: I1203 02:24:18.572085 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:24:18 crc kubenswrapper[4912]: E1203 02:24:18.573091 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:24:31 crc kubenswrapper[4912]: I1203 02:24:31.571918 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:24:31 crc kubenswrapper[4912]: E1203 02:24:31.572612 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:24:44 crc kubenswrapper[4912]: I1203 02:24:44.573230 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:24:44 crc kubenswrapper[4912]: E1203 02:24:44.576556 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:24:55 crc kubenswrapper[4912]: I1203 02:24:55.571637 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:24:55 crc kubenswrapper[4912]: E1203 02:24:55.572401 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:25:09 crc kubenswrapper[4912]: I1203 02:25:09.571507 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:25:09 crc kubenswrapper[4912]: E1203 02:25:09.573122 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:25:24 crc kubenswrapper[4912]: I1203 02:25:24.591493 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:25:24 crc kubenswrapper[4912]: E1203 02:25:24.592964 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.940109 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m6588"] Dec 03 02:25:33 crc kubenswrapper[4912]: E1203 02:25:33.941312 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerName="registry-server" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.941333 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerName="registry-server" Dec 03 02:25:33 crc kubenswrapper[4912]: E1203 02:25:33.941357 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerName="extract-content" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.941366 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerName="extract-content" Dec 03 02:25:33 crc kubenswrapper[4912]: E1203 02:25:33.941386 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerName="extract-utilities" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.941395 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerName="extract-utilities" Dec 03 02:25:33 crc kubenswrapper[4912]: E1203 02:25:33.941421 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerName="extract-content" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.941453 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerName="extract-content" Dec 03 02:25:33 crc kubenswrapper[4912]: E1203 02:25:33.941474 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerName="extract-utilities" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.941482 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerName="extract-utilities" Dec 03 02:25:33 crc kubenswrapper[4912]: E1203 02:25:33.941527 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerName="registry-server" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.941538 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerName="registry-server" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.941940 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="3266b492-f2dc-42a4-86a5-1d52438532b8" containerName="registry-server" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.941962 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc7b7639-cebd-4b44-b0fe-5cffdcc227d5" containerName="registry-server" Dec 03 02:25:33 crc kubenswrapper[4912]: I1203 02:25:33.964370 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.002993 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6588"] Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.145512 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-utilities\") pod \"redhat-marketplace-m6588\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.145641 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc24w\" (UniqueName: \"kubernetes.io/projected/b11c0419-2568-4e2e-98d4-b56e3f8866e5-kube-api-access-nc24w\") pod \"redhat-marketplace-m6588\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.145828 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-catalog-content\") pod \"redhat-marketplace-m6588\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.247385 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc24w\" (UniqueName: \"kubernetes.io/projected/b11c0419-2568-4e2e-98d4-b56e3f8866e5-kube-api-access-nc24w\") pod \"redhat-marketplace-m6588\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.247543 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-catalog-content\") pod \"redhat-marketplace-m6588\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.247626 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-utilities\") pod \"redhat-marketplace-m6588\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.248097 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-catalog-content\") pod \"redhat-marketplace-m6588\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.248251 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-utilities\") pod \"redhat-marketplace-m6588\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.267353 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc24w\" (UniqueName: \"kubernetes.io/projected/b11c0419-2568-4e2e-98d4-b56e3f8866e5-kube-api-access-nc24w\") pod \"redhat-marketplace-m6588\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.321229 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:34 crc kubenswrapper[4912]: I1203 02:25:34.850332 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6588"] Dec 03 02:25:35 crc kubenswrapper[4912]: I1203 02:25:35.049059 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6588" event={"ID":"b11c0419-2568-4e2e-98d4-b56e3f8866e5","Type":"ContainerStarted","Data":"764b3bf482c93770da7cc74e76ddb0b49bf5ac78a8bfcf4cdc5822e7f1b09a17"} Dec 03 02:25:35 crc kubenswrapper[4912]: I1203 02:25:35.571933 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:25:35 crc kubenswrapper[4912]: E1203 02:25:35.572313 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:25:36 crc kubenswrapper[4912]: I1203 02:25:36.059713 4912 generic.go:334] "Generic (PLEG): container finished" podID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerID="1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285" exitCode=0 Dec 03 02:25:36 crc kubenswrapper[4912]: I1203 02:25:36.059757 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6588" event={"ID":"b11c0419-2568-4e2e-98d4-b56e3f8866e5","Type":"ContainerDied","Data":"1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285"} Dec 03 02:25:36 crc kubenswrapper[4912]: I1203 02:25:36.062734 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 02:25:37 crc kubenswrapper[4912]: I1203 02:25:37.073204 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6588" event={"ID":"b11c0419-2568-4e2e-98d4-b56e3f8866e5","Type":"ContainerStarted","Data":"ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92"} Dec 03 02:25:38 crc kubenswrapper[4912]: I1203 02:25:38.088455 4912 generic.go:334] "Generic (PLEG): container finished" podID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerID="ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92" exitCode=0 Dec 03 02:25:38 crc kubenswrapper[4912]: I1203 02:25:38.088563 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6588" event={"ID":"b11c0419-2568-4e2e-98d4-b56e3f8866e5","Type":"ContainerDied","Data":"ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92"} Dec 03 02:25:39 crc kubenswrapper[4912]: I1203 02:25:39.107363 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6588" event={"ID":"b11c0419-2568-4e2e-98d4-b56e3f8866e5","Type":"ContainerStarted","Data":"323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d"} Dec 03 02:25:39 crc kubenswrapper[4912]: I1203 02:25:39.149988 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m6588" podStartSLOduration=3.630813104 podStartE2EDuration="6.149956464s" podCreationTimestamp="2025-12-03 02:25:33 +0000 UTC" firstStartedPulling="2025-12-03 02:25:36.062530271 +0000 UTC m=+7321.704550831" lastFinishedPulling="2025-12-03 02:25:38.581673621 +0000 UTC m=+7324.223694191" observedRunningTime="2025-12-03 02:25:39.129816434 +0000 UTC m=+7324.771837004" watchObservedRunningTime="2025-12-03 02:25:39.149956464 +0000 UTC m=+7324.791977064" Dec 03 02:25:44 crc kubenswrapper[4912]: I1203 02:25:44.321727 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:44 crc kubenswrapper[4912]: I1203 02:25:44.322498 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:44 crc kubenswrapper[4912]: I1203 02:25:44.375536 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:45 crc kubenswrapper[4912]: I1203 02:25:45.255084 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:45 crc kubenswrapper[4912]: I1203 02:25:45.337963 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6588"] Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.208502 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m6588" podUID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerName="registry-server" containerID="cri-o://323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d" gracePeriod=2 Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.572222 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:25:47 crc kubenswrapper[4912]: E1203 02:25:47.572975 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.721690 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.918269 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-catalog-content\") pod \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.918375 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-utilities\") pod \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.918608 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc24w\" (UniqueName: \"kubernetes.io/projected/b11c0419-2568-4e2e-98d4-b56e3f8866e5-kube-api-access-nc24w\") pod \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\" (UID: \"b11c0419-2568-4e2e-98d4-b56e3f8866e5\") " Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.919298 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-utilities" (OuterVolumeSpecName: "utilities") pod "b11c0419-2568-4e2e-98d4-b56e3f8866e5" (UID: "b11c0419-2568-4e2e-98d4-b56e3f8866e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.919907 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.929004 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11c0419-2568-4e2e-98d4-b56e3f8866e5-kube-api-access-nc24w" (OuterVolumeSpecName: "kube-api-access-nc24w") pod "b11c0419-2568-4e2e-98d4-b56e3f8866e5" (UID: "b11c0419-2568-4e2e-98d4-b56e3f8866e5"). InnerVolumeSpecName "kube-api-access-nc24w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:25:47 crc kubenswrapper[4912]: I1203 02:25:47.936359 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11c0419-2568-4e2e-98d4-b56e3f8866e5" (UID: "b11c0419-2568-4e2e-98d4-b56e3f8866e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.021611 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11c0419-2568-4e2e-98d4-b56e3f8866e5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.021851 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc24w\" (UniqueName: \"kubernetes.io/projected/b11c0419-2568-4e2e-98d4-b56e3f8866e5-kube-api-access-nc24w\") on node \"crc\" DevicePath \"\"" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.221232 4912 generic.go:334] "Generic (PLEG): container finished" podID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerID="323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d" exitCode=0 Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.221389 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m6588" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.221403 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6588" event={"ID":"b11c0419-2568-4e2e-98d4-b56e3f8866e5","Type":"ContainerDied","Data":"323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d"} Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.222522 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m6588" event={"ID":"b11c0419-2568-4e2e-98d4-b56e3f8866e5","Type":"ContainerDied","Data":"764b3bf482c93770da7cc74e76ddb0b49bf5ac78a8bfcf4cdc5822e7f1b09a17"} Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.222550 4912 scope.go:117] "RemoveContainer" containerID="323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.258077 4912 scope.go:117] "RemoveContainer" containerID="ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.263521 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6588"] Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.289387 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m6588"] Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.318166 4912 scope.go:117] "RemoveContainer" containerID="1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.343845 4912 scope.go:117] "RemoveContainer" containerID="323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d" Dec 03 02:25:48 crc kubenswrapper[4912]: E1203 02:25:48.344362 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d\": container with ID starting with 323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d not found: ID does not exist" containerID="323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.344404 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d"} err="failed to get container status \"323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d\": rpc error: code = NotFound desc = could not find container \"323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d\": container with ID starting with 323a0200ea78b1bbd321902486a1a8f8d8738ba82fb1301147cd9c9dffb1de0d not found: ID does not exist" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.344443 4912 scope.go:117] "RemoveContainer" containerID="ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92" Dec 03 02:25:48 crc kubenswrapper[4912]: E1203 02:25:48.344781 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92\": container with ID starting with ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92 not found: ID does not exist" containerID="ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.344812 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92"} err="failed to get container status \"ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92\": rpc error: code = NotFound desc = could not find container \"ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92\": container with ID starting with ecd2bc7fc6b82cbfabdf756117bf27827d60f3b98bd72a832f39add1778e5f92 not found: ID does not exist" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.344834 4912 scope.go:117] "RemoveContainer" containerID="1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285" Dec 03 02:25:48 crc kubenswrapper[4912]: E1203 02:25:48.345066 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285\": container with ID starting with 1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285 not found: ID does not exist" containerID="1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.345104 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285"} err="failed to get container status \"1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285\": rpc error: code = NotFound desc = could not find container \"1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285\": container with ID starting with 1a79989e8eada27775f5d8fb2111884298ca5fa02290dc9711122bca90834285 not found: ID does not exist" Dec 03 02:25:48 crc kubenswrapper[4912]: I1203 02:25:48.596784 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" path="/var/lib/kubelet/pods/b11c0419-2568-4e2e-98d4-b56e3f8866e5/volumes" Dec 03 02:26:01 crc kubenswrapper[4912]: I1203 02:26:01.572495 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:26:01 crc kubenswrapper[4912]: E1203 02:26:01.573672 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:26:16 crc kubenswrapper[4912]: I1203 02:26:16.573197 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:26:16 crc kubenswrapper[4912]: E1203 02:26:16.574253 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:26:29 crc kubenswrapper[4912]: I1203 02:26:29.573007 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:26:30 crc kubenswrapper[4912]: I1203 02:26:30.800089 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"8d0c2b90c0689a7fde786f8759ce917ec7c2d8c36cbf01ee538d42fc7f0328a7"} Dec 03 02:28:48 crc kubenswrapper[4912]: I1203 02:28:48.077473 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:28:48 crc kubenswrapper[4912]: I1203 02:28:48.078013 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:29:18 crc kubenswrapper[4912]: I1203 02:29:18.077460 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:29:18 crc kubenswrapper[4912]: I1203 02:29:18.078031 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.207285 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cb2zw"] Dec 03 02:29:33 crc kubenswrapper[4912]: E1203 02:29:33.208625 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerName="registry-server" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.208649 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerName="registry-server" Dec 03 02:29:33 crc kubenswrapper[4912]: E1203 02:29:33.208725 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerName="extract-utilities" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.208739 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerName="extract-utilities" Dec 03 02:29:33 crc kubenswrapper[4912]: E1203 02:29:33.208767 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerName="extract-content" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.208783 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerName="extract-content" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.209659 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11c0419-2568-4e2e-98d4-b56e3f8866e5" containerName="registry-server" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.213080 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.242153 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cb2zw"] Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.312690 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljmwz\" (UniqueName: \"kubernetes.io/projected/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-kube-api-access-ljmwz\") pod \"certified-operators-cb2zw\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.312933 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-catalog-content\") pod \"certified-operators-cb2zw\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.313268 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-utilities\") pod \"certified-operators-cb2zw\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.414646 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-utilities\") pod \"certified-operators-cb2zw\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.414752 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljmwz\" (UniqueName: \"kubernetes.io/projected/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-kube-api-access-ljmwz\") pod \"certified-operators-cb2zw\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.414910 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-catalog-content\") pod \"certified-operators-cb2zw\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.415614 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-catalog-content\") pod \"certified-operators-cb2zw\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.415995 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-utilities\") pod \"certified-operators-cb2zw\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.449627 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljmwz\" (UniqueName: \"kubernetes.io/projected/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-kube-api-access-ljmwz\") pod \"certified-operators-cb2zw\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:33 crc kubenswrapper[4912]: I1203 02:29:33.541184 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:34 crc kubenswrapper[4912]: I1203 02:29:34.110673 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cb2zw"] Dec 03 02:29:34 crc kubenswrapper[4912]: I1203 02:29:34.412056 4912 generic.go:334] "Generic (PLEG): container finished" podID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerID="965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e" exitCode=0 Dec 03 02:29:34 crc kubenswrapper[4912]: I1203 02:29:34.412114 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2zw" event={"ID":"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c","Type":"ContainerDied","Data":"965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e"} Dec 03 02:29:34 crc kubenswrapper[4912]: I1203 02:29:34.412343 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2zw" event={"ID":"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c","Type":"ContainerStarted","Data":"59f60bc21e3c0e4215ab820f778176d6f24a7290496bb5c0c327c8bb9d3fa7c3"} Dec 03 02:29:35 crc kubenswrapper[4912]: I1203 02:29:35.435840 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2zw" event={"ID":"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c","Type":"ContainerStarted","Data":"ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e"} Dec 03 02:29:36 crc kubenswrapper[4912]: I1203 02:29:36.448784 4912 generic.go:334] "Generic (PLEG): container finished" podID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerID="ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e" exitCode=0 Dec 03 02:29:36 crc kubenswrapper[4912]: I1203 02:29:36.448830 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2zw" event={"ID":"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c","Type":"ContainerDied","Data":"ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e"} Dec 03 02:29:37 crc kubenswrapper[4912]: I1203 02:29:37.461097 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2zw" event={"ID":"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c","Type":"ContainerStarted","Data":"9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2"} Dec 03 02:29:37 crc kubenswrapper[4912]: I1203 02:29:37.477244 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cb2zw" podStartSLOduration=2.009476014 podStartE2EDuration="4.477224588s" podCreationTimestamp="2025-12-03 02:29:33 +0000 UTC" firstStartedPulling="2025-12-03 02:29:34.414196116 +0000 UTC m=+7560.056216676" lastFinishedPulling="2025-12-03 02:29:36.88194468 +0000 UTC m=+7562.523965250" observedRunningTime="2025-12-03 02:29:37.474691468 +0000 UTC m=+7563.116712038" watchObservedRunningTime="2025-12-03 02:29:37.477224588 +0000 UTC m=+7563.119245148" Dec 03 02:29:43 crc kubenswrapper[4912]: I1203 02:29:43.543492 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:43 crc kubenswrapper[4912]: I1203 02:29:43.544164 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:43 crc kubenswrapper[4912]: I1203 02:29:43.612054 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:44 crc kubenswrapper[4912]: I1203 02:29:44.619802 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:44 crc kubenswrapper[4912]: I1203 02:29:44.687398 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cb2zw"] Dec 03 02:29:46 crc kubenswrapper[4912]: I1203 02:29:46.556614 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cb2zw" podUID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerName="registry-server" containerID="cri-o://9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2" gracePeriod=2 Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.125044 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.249203 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljmwz\" (UniqueName: \"kubernetes.io/projected/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-kube-api-access-ljmwz\") pod \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.249514 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-utilities\") pod \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.249756 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-catalog-content\") pod \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\" (UID: \"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c\") " Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.250406 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-utilities" (OuterVolumeSpecName: "utilities") pod "2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" (UID: "2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.251090 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.258148 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-kube-api-access-ljmwz" (OuterVolumeSpecName: "kube-api-access-ljmwz") pod "2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" (UID: "2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c"). InnerVolumeSpecName "kube-api-access-ljmwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.299947 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" (UID: "2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.365479 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.365515 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljmwz\" (UniqueName: \"kubernetes.io/projected/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c-kube-api-access-ljmwz\") on node \"crc\" DevicePath \"\"" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.573735 4912 generic.go:334] "Generic (PLEG): container finished" podID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerID="9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2" exitCode=0 Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.575119 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cb2zw" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.575138 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2zw" event={"ID":"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c","Type":"ContainerDied","Data":"9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2"} Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.583958 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cb2zw" event={"ID":"2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c","Type":"ContainerDied","Data":"59f60bc21e3c0e4215ab820f778176d6f24a7290496bb5c0c327c8bb9d3fa7c3"} Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.584012 4912 scope.go:117] "RemoveContainer" containerID="9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.625001 4912 scope.go:117] "RemoveContainer" containerID="ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.681417 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cb2zw"] Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.689956 4912 scope.go:117] "RemoveContainer" containerID="965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.704143 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cb2zw"] Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.714719 4912 scope.go:117] "RemoveContainer" containerID="9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2" Dec 03 02:29:47 crc kubenswrapper[4912]: E1203 02:29:47.715313 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2\": container with ID starting with 9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2 not found: ID does not exist" containerID="9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.715368 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2"} err="failed to get container status \"9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2\": rpc error: code = NotFound desc = could not find container \"9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2\": container with ID starting with 9349e9b7898c85f1d2ec9b39c3a5ea64d553fe34b32bf156d745837c5257dfd2 not found: ID does not exist" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.715400 4912 scope.go:117] "RemoveContainer" containerID="ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e" Dec 03 02:29:47 crc kubenswrapper[4912]: E1203 02:29:47.715761 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e\": container with ID starting with ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e not found: ID does not exist" containerID="ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.715808 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e"} err="failed to get container status \"ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e\": rpc error: code = NotFound desc = could not find container \"ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e\": container with ID starting with ffa33d2058a2ad76dbfd620d6edb41639d6bf5fbe0bb1b0256389ae0a76bd33e not found: ID does not exist" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.715836 4912 scope.go:117] "RemoveContainer" containerID="965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e" Dec 03 02:29:47 crc kubenswrapper[4912]: E1203 02:29:47.716097 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e\": container with ID starting with 965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e not found: ID does not exist" containerID="965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e" Dec 03 02:29:47 crc kubenswrapper[4912]: I1203 02:29:47.716128 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e"} err="failed to get container status \"965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e\": rpc error: code = NotFound desc = could not find container \"965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e\": container with ID starting with 965412b3df94c7f2a71f4dda28b00350c19a2ece17f935a158f33a77ad739c0e not found: ID does not exist" Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.078264 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.078325 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.078371 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.079179 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d0c2b90c0689a7fde786f8759ce917ec7c2d8c36cbf01ee538d42fc7f0328a7"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.079250 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://8d0c2b90c0689a7fde786f8759ce917ec7c2d8c36cbf01ee538d42fc7f0328a7" gracePeriod=600 Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.590512 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" path="/var/lib/kubelet/pods/2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c/volumes" Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.603662 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="8d0c2b90c0689a7fde786f8759ce917ec7c2d8c36cbf01ee538d42fc7f0328a7" exitCode=0 Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.603711 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"8d0c2b90c0689a7fde786f8759ce917ec7c2d8c36cbf01ee538d42fc7f0328a7"} Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.603747 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb"} Dec 03 02:29:48 crc kubenswrapper[4912]: I1203 02:29:48.603763 4912 scope.go:117] "RemoveContainer" containerID="b068fa43e14e78e8435f2af30597c3180d954b258ed492c35d039adb48e9aba5" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.207115 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p"] Dec 03 02:30:00 crc kubenswrapper[4912]: E1203 02:30:00.208267 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerName="extract-content" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.208285 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerName="extract-content" Dec 03 02:30:00 crc kubenswrapper[4912]: E1203 02:30:00.208325 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerName="extract-utilities" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.208333 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerName="extract-utilities" Dec 03 02:30:00 crc kubenswrapper[4912]: E1203 02:30:00.208382 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerName="registry-server" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.208389 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerName="registry-server" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.208657 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a8fe6ac-37d1-4eec-b5e2-de74dbf1165c" containerName="registry-server" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.209441 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.223287 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p"] Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.229470 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.244572 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.390713 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-config-volume\") pod \"collect-profiles-29412150-r5t2p\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.390870 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-secret-volume\") pod \"collect-profiles-29412150-r5t2p\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.390922 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glp56\" (UniqueName: \"kubernetes.io/projected/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-kube-api-access-glp56\") pod \"collect-profiles-29412150-r5t2p\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.497696 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-secret-volume\") pod \"collect-profiles-29412150-r5t2p\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.497772 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glp56\" (UniqueName: \"kubernetes.io/projected/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-kube-api-access-glp56\") pod \"collect-profiles-29412150-r5t2p\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.497851 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-config-volume\") pod \"collect-profiles-29412150-r5t2p\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.498756 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-config-volume\") pod \"collect-profiles-29412150-r5t2p\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.518405 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-secret-volume\") pod \"collect-profiles-29412150-r5t2p\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.523476 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glp56\" (UniqueName: \"kubernetes.io/projected/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-kube-api-access-glp56\") pod \"collect-profiles-29412150-r5t2p\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:00 crc kubenswrapper[4912]: I1203 02:30:00.601697 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:01 crc kubenswrapper[4912]: W1203 02:30:01.136722 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2547f1fe_6c1c_439b_b7c0_b7ebd5335e25.slice/crio-37d37bc5fc3fa2ab16d6b542fae4e6395244a5103d349032c4ca8a34e461d1d6 WatchSource:0}: Error finding container 37d37bc5fc3fa2ab16d6b542fae4e6395244a5103d349032c4ca8a34e461d1d6: Status 404 returned error can't find the container with id 37d37bc5fc3fa2ab16d6b542fae4e6395244a5103d349032c4ca8a34e461d1d6 Dec 03 02:30:01 crc kubenswrapper[4912]: I1203 02:30:01.139152 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p"] Dec 03 02:30:01 crc kubenswrapper[4912]: I1203 02:30:01.782774 4912 generic.go:334] "Generic (PLEG): container finished" podID="2547f1fe-6c1c-439b-b7c0-b7ebd5335e25" containerID="d6ddeb852d49ab19cbb766f89c00f778b3fd82c280799862ed638b7faf063f48" exitCode=0 Dec 03 02:30:01 crc kubenswrapper[4912]: I1203 02:30:01.782850 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" event={"ID":"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25","Type":"ContainerDied","Data":"d6ddeb852d49ab19cbb766f89c00f778b3fd82c280799862ed638b7faf063f48"} Dec 03 02:30:01 crc kubenswrapper[4912]: I1203 02:30:01.783132 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" event={"ID":"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25","Type":"ContainerStarted","Data":"37d37bc5fc3fa2ab16d6b542fae4e6395244a5103d349032c4ca8a34e461d1d6"} Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.309415 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.466461 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glp56\" (UniqueName: \"kubernetes.io/projected/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-kube-api-access-glp56\") pod \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.466588 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-config-volume\") pod \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.466776 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-secret-volume\") pod \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\" (UID: \"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25\") " Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.467403 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-config-volume" (OuterVolumeSpecName: "config-volume") pod "2547f1fe-6c1c-439b-b7c0-b7ebd5335e25" (UID: "2547f1fe-6c1c-439b-b7c0-b7ebd5335e25"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.468611 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.472577 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-kube-api-access-glp56" (OuterVolumeSpecName: "kube-api-access-glp56") pod "2547f1fe-6c1c-439b-b7c0-b7ebd5335e25" (UID: "2547f1fe-6c1c-439b-b7c0-b7ebd5335e25"). InnerVolumeSpecName "kube-api-access-glp56". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.472907 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2547f1fe-6c1c-439b-b7c0-b7ebd5335e25" (UID: "2547f1fe-6c1c-439b-b7c0-b7ebd5335e25"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.571222 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.571263 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glp56\" (UniqueName: \"kubernetes.io/projected/2547f1fe-6c1c-439b-b7c0-b7ebd5335e25-kube-api-access-glp56\") on node \"crc\" DevicePath \"\"" Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.808785 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" event={"ID":"2547f1fe-6c1c-439b-b7c0-b7ebd5335e25","Type":"ContainerDied","Data":"37d37bc5fc3fa2ab16d6b542fae4e6395244a5103d349032c4ca8a34e461d1d6"} Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.809046 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37d37bc5fc3fa2ab16d6b542fae4e6395244a5103d349032c4ca8a34e461d1d6" Dec 03 02:30:03 crc kubenswrapper[4912]: I1203 02:30:03.808870 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412150-r5t2p" Dec 03 02:30:04 crc kubenswrapper[4912]: I1203 02:30:04.475095 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf"] Dec 03 02:30:04 crc kubenswrapper[4912]: I1203 02:30:04.488879 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412105-jbkkf"] Dec 03 02:30:04 crc kubenswrapper[4912]: I1203 02:30:04.655479 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dff648e-50e3-433c-b9e6-d3a83c0ef3aa" path="/var/lib/kubelet/pods/9dff648e-50e3-433c-b9e6-d3a83c0ef3aa/volumes" Dec 03 02:30:29 crc kubenswrapper[4912]: I1203 02:30:29.169244 4912 scope.go:117] "RemoveContainer" containerID="c114602a364bed6f635a573ee34f2bb3490154855fa8b6090cd3adf54bf8e76a" Dec 03 02:31:48 crc kubenswrapper[4912]: I1203 02:31:48.078092 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:31:48 crc kubenswrapper[4912]: I1203 02:31:48.078812 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.362221 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mgm9z"] Dec 03 02:32:02 crc kubenswrapper[4912]: E1203 02:32:02.363126 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2547f1fe-6c1c-439b-b7c0-b7ebd5335e25" containerName="collect-profiles" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.363139 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2547f1fe-6c1c-439b-b7c0-b7ebd5335e25" containerName="collect-profiles" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.363725 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2547f1fe-6c1c-439b-b7c0-b7ebd5335e25" containerName="collect-profiles" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.365344 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.379932 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mgm9z"] Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.504221 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-catalog-content\") pod \"redhat-operators-mgm9z\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.504412 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-utilities\") pod \"redhat-operators-mgm9z\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.504517 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-528l4\" (UniqueName: \"kubernetes.io/projected/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-kube-api-access-528l4\") pod \"redhat-operators-mgm9z\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.606870 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-utilities\") pod \"redhat-operators-mgm9z\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.607226 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-528l4\" (UniqueName: \"kubernetes.io/projected/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-kube-api-access-528l4\") pod \"redhat-operators-mgm9z\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.607409 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-catalog-content\") pod \"redhat-operators-mgm9z\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.607790 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-utilities\") pod \"redhat-operators-mgm9z\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.608297 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-catalog-content\") pod \"redhat-operators-mgm9z\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.630517 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-528l4\" (UniqueName: \"kubernetes.io/projected/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-kube-api-access-528l4\") pod \"redhat-operators-mgm9z\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:02 crc kubenswrapper[4912]: I1203 02:32:02.730638 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:03 crc kubenswrapper[4912]: I1203 02:32:03.308506 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mgm9z"] Dec 03 02:32:03 crc kubenswrapper[4912]: I1203 02:32:03.502124 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgm9z" event={"ID":"caa09c0b-cb22-4c5c-a3a1-0269f51975ef","Type":"ContainerStarted","Data":"0dc5a7dd91d6a363f57b4fd4d7e0879bb83f0ee4c2d1c61ee5605690b0060e5e"} Dec 03 02:32:04 crc kubenswrapper[4912]: I1203 02:32:04.515340 4912 generic.go:334] "Generic (PLEG): container finished" podID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerID="55e4eeafb54d38486052f752555e327363950ca380b81c7d5fa9202362ddcebd" exitCode=0 Dec 03 02:32:04 crc kubenswrapper[4912]: I1203 02:32:04.515471 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgm9z" event={"ID":"caa09c0b-cb22-4c5c-a3a1-0269f51975ef","Type":"ContainerDied","Data":"55e4eeafb54d38486052f752555e327363950ca380b81c7d5fa9202362ddcebd"} Dec 03 02:32:04 crc kubenswrapper[4912]: I1203 02:32:04.517966 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 02:32:05 crc kubenswrapper[4912]: I1203 02:32:05.527076 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgm9z" event={"ID":"caa09c0b-cb22-4c5c-a3a1-0269f51975ef","Type":"ContainerStarted","Data":"7b86baefc423c31ea916931508dba7de3292d139fb7a90441cd7da0c8fc1fb81"} Dec 03 02:32:08 crc kubenswrapper[4912]: I1203 02:32:08.575993 4912 generic.go:334] "Generic (PLEG): container finished" podID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerID="7b86baefc423c31ea916931508dba7de3292d139fb7a90441cd7da0c8fc1fb81" exitCode=0 Dec 03 02:32:08 crc kubenswrapper[4912]: I1203 02:32:08.592493 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgm9z" event={"ID":"caa09c0b-cb22-4c5c-a3a1-0269f51975ef","Type":"ContainerDied","Data":"7b86baefc423c31ea916931508dba7de3292d139fb7a90441cd7da0c8fc1fb81"} Dec 03 02:32:09 crc kubenswrapper[4912]: I1203 02:32:09.598315 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgm9z" event={"ID":"caa09c0b-cb22-4c5c-a3a1-0269f51975ef","Type":"ContainerStarted","Data":"6a3e1d78be372dc7650e89491772f673e3e667cee2c3ffa3427cd5e87e05ff53"} Dec 03 02:32:09 crc kubenswrapper[4912]: I1203 02:32:09.626668 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mgm9z" podStartSLOduration=3.054343284 podStartE2EDuration="7.626639685s" podCreationTimestamp="2025-12-03 02:32:02 +0000 UTC" firstStartedPulling="2025-12-03 02:32:04.517726861 +0000 UTC m=+7710.159747421" lastFinishedPulling="2025-12-03 02:32:09.090023222 +0000 UTC m=+7714.732043822" observedRunningTime="2025-12-03 02:32:09.615184507 +0000 UTC m=+7715.257205087" watchObservedRunningTime="2025-12-03 02:32:09.626639685 +0000 UTC m=+7715.268660235" Dec 03 02:32:12 crc kubenswrapper[4912]: I1203 02:32:12.731130 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:12 crc kubenswrapper[4912]: I1203 02:32:12.731632 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:13 crc kubenswrapper[4912]: I1203 02:32:13.788406 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mgm9z" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerName="registry-server" probeResult="failure" output=< Dec 03 02:32:13 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 02:32:13 crc kubenswrapper[4912]: > Dec 03 02:32:18 crc kubenswrapper[4912]: I1203 02:32:18.078511 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:32:18 crc kubenswrapper[4912]: I1203 02:32:18.079487 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:32:22 crc kubenswrapper[4912]: I1203 02:32:22.801676 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:22 crc kubenswrapper[4912]: I1203 02:32:22.871387 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:23 crc kubenswrapper[4912]: I1203 02:32:23.043146 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mgm9z"] Dec 03 02:32:24 crc kubenswrapper[4912]: I1203 02:32:24.786254 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mgm9z" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerName="registry-server" containerID="cri-o://6a3e1d78be372dc7650e89491772f673e3e667cee2c3ffa3427cd5e87e05ff53" gracePeriod=2 Dec 03 02:32:24 crc kubenswrapper[4912]: E1203 02:32:24.989121 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcaa09c0b_cb22_4c5c_a3a1_0269f51975ef.slice/crio-6a3e1d78be372dc7650e89491772f673e3e667cee2c3ffa3427cd5e87e05ff53.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcaa09c0b_cb22_4c5c_a3a1_0269f51975ef.slice/crio-conmon-6a3e1d78be372dc7650e89491772f673e3e667cee2c3ffa3427cd5e87e05ff53.scope\": RecentStats: unable to find data in memory cache]" Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.809482 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgm9z" event={"ID":"caa09c0b-cb22-4c5c-a3a1-0269f51975ef","Type":"ContainerDied","Data":"6a3e1d78be372dc7650e89491772f673e3e667cee2c3ffa3427cd5e87e05ff53"} Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.809411 4912 generic.go:334] "Generic (PLEG): container finished" podID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerID="6a3e1d78be372dc7650e89491772f673e3e667cee2c3ffa3427cd5e87e05ff53" exitCode=0 Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.810258 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mgm9z" event={"ID":"caa09c0b-cb22-4c5c-a3a1-0269f51975ef","Type":"ContainerDied","Data":"0dc5a7dd91d6a363f57b4fd4d7e0879bb83f0ee4c2d1c61ee5605690b0060e5e"} Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.810284 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dc5a7dd91d6a363f57b4fd4d7e0879bb83f0ee4c2d1c61ee5605690b0060e5e" Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.911138 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.981971 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-catalog-content\") pod \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.983452 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-528l4\" (UniqueName: \"kubernetes.io/projected/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-kube-api-access-528l4\") pod \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.983645 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-utilities\") pod \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\" (UID: \"caa09c0b-cb22-4c5c-a3a1-0269f51975ef\") " Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.986451 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-utilities" (OuterVolumeSpecName: "utilities") pod "caa09c0b-cb22-4c5c-a3a1-0269f51975ef" (UID: "caa09c0b-cb22-4c5c-a3a1-0269f51975ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:32:25 crc kubenswrapper[4912]: I1203 02:32:25.995394 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-kube-api-access-528l4" (OuterVolumeSpecName: "kube-api-access-528l4") pod "caa09c0b-cb22-4c5c-a3a1-0269f51975ef" (UID: "caa09c0b-cb22-4c5c-a3a1-0269f51975ef"). InnerVolumeSpecName "kube-api-access-528l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:32:26 crc kubenswrapper[4912]: I1203 02:32:26.087090 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-528l4\" (UniqueName: \"kubernetes.io/projected/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-kube-api-access-528l4\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:26 crc kubenswrapper[4912]: I1203 02:32:26.087122 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:26 crc kubenswrapper[4912]: I1203 02:32:26.087667 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "caa09c0b-cb22-4c5c-a3a1-0269f51975ef" (UID: "caa09c0b-cb22-4c5c-a3a1-0269f51975ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:32:26 crc kubenswrapper[4912]: I1203 02:32:26.189640 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/caa09c0b-cb22-4c5c-a3a1-0269f51975ef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:26 crc kubenswrapper[4912]: I1203 02:32:26.818676 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mgm9z" Dec 03 02:32:26 crc kubenswrapper[4912]: I1203 02:32:26.845132 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mgm9z"] Dec 03 02:32:26 crc kubenswrapper[4912]: I1203 02:32:26.855342 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mgm9z"] Dec 03 02:32:28 crc kubenswrapper[4912]: I1203 02:32:28.587514 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" path="/var/lib/kubelet/pods/caa09c0b-cb22-4c5c-a3a1-0269f51975ef/volumes" Dec 03 02:32:48 crc kubenswrapper[4912]: I1203 02:32:48.077820 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:32:48 crc kubenswrapper[4912]: I1203 02:32:48.078641 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:32:48 crc kubenswrapper[4912]: I1203 02:32:48.078713 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:32:48 crc kubenswrapper[4912]: I1203 02:32:48.080343 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:32:48 crc kubenswrapper[4912]: I1203 02:32:48.080412 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" gracePeriod=600 Dec 03 02:32:48 crc kubenswrapper[4912]: E1203 02:32:48.221026 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:32:49 crc kubenswrapper[4912]: I1203 02:32:49.119997 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" exitCode=0 Dec 03 02:32:49 crc kubenswrapper[4912]: I1203 02:32:49.120098 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb"} Dec 03 02:32:49 crc kubenswrapper[4912]: I1203 02:32:49.120503 4912 scope.go:117] "RemoveContainer" containerID="8d0c2b90c0689a7fde786f8759ce917ec7c2d8c36cbf01ee538d42fc7f0328a7" Dec 03 02:32:49 crc kubenswrapper[4912]: I1203 02:32:49.121708 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:32:49 crc kubenswrapper[4912]: E1203 02:32:49.122713 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:32:54 crc kubenswrapper[4912]: I1203 02:32:54.207072 4912 generic.go:334] "Generic (PLEG): container finished" podID="d5f2ea7f-cdc9-4573-8b16-6561c761f91f" containerID="95ca4f945c84b30eb7280d9304ad12914e810485b59bd9ae9bdd4640819db579" exitCode=0 Dec 03 02:32:54 crc kubenswrapper[4912]: I1203 02:32:54.207126 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d5f2ea7f-cdc9-4573-8b16-6561c761f91f","Type":"ContainerDied","Data":"95ca4f945c84b30eb7280d9304ad12914e810485b59bd9ae9bdd4640819db579"} Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.726935 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.839592 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-workdir\") pod \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.839670 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-temporary\") pod \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.839733 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ssh-key\") pod \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.839801 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.840094 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw9b2\" (UniqueName: \"kubernetes.io/projected/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-kube-api-access-zw9b2\") pod \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.840143 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-config-data\") pod \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.840325 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config-secret\") pod \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.840550 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "d5f2ea7f-cdc9-4573-8b16-6561c761f91f" (UID: "d5f2ea7f-cdc9-4573-8b16-6561c761f91f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.841456 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config\") pod \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.841497 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ca-certs\") pod \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\" (UID: \"d5f2ea7f-cdc9-4573-8b16-6561c761f91f\") " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.841706 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-config-data" (OuterVolumeSpecName: "config-data") pod "d5f2ea7f-cdc9-4573-8b16-6561c761f91f" (UID: "d5f2ea7f-cdc9-4573-8b16-6561c761f91f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.842330 4912 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.842351 4912 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.845698 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-kube-api-access-zw9b2" (OuterVolumeSpecName: "kube-api-access-zw9b2") pod "d5f2ea7f-cdc9-4573-8b16-6561c761f91f" (UID: "d5f2ea7f-cdc9-4573-8b16-6561c761f91f"). InnerVolumeSpecName "kube-api-access-zw9b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.847761 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "d5f2ea7f-cdc9-4573-8b16-6561c761f91f" (UID: "d5f2ea7f-cdc9-4573-8b16-6561c761f91f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.850192 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "d5f2ea7f-cdc9-4573-8b16-6561c761f91f" (UID: "d5f2ea7f-cdc9-4573-8b16-6561c761f91f"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.874189 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d5f2ea7f-cdc9-4573-8b16-6561c761f91f" (UID: "d5f2ea7f-cdc9-4573-8b16-6561c761f91f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.881078 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "d5f2ea7f-cdc9-4573-8b16-6561c761f91f" (UID: "d5f2ea7f-cdc9-4573-8b16-6561c761f91f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.889556 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d5f2ea7f-cdc9-4573-8b16-6561c761f91f" (UID: "d5f2ea7f-cdc9-4573-8b16-6561c761f91f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.912427 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d5f2ea7f-cdc9-4573-8b16-6561c761f91f" (UID: "d5f2ea7f-cdc9-4573-8b16-6561c761f91f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.944269 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.944317 4912 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.944334 4912 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.944348 4912 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.944362 4912 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.945000 4912 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.945027 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw9b2\" (UniqueName: \"kubernetes.io/projected/d5f2ea7f-cdc9-4573-8b16-6561c761f91f-kube-api-access-zw9b2\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:55 crc kubenswrapper[4912]: I1203 02:32:55.975686 4912 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 03 02:32:56 crc kubenswrapper[4912]: I1203 02:32:56.047446 4912 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 03 02:32:56 crc kubenswrapper[4912]: I1203 02:32:56.234949 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d5f2ea7f-cdc9-4573-8b16-6561c761f91f","Type":"ContainerDied","Data":"3dc0ad0e970fcf78de2548b555d30324f15505274865780bdafb05a564b5db28"} Dec 03 02:32:56 crc kubenswrapper[4912]: I1203 02:32:56.234993 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dc0ad0e970fcf78de2548b555d30324f15505274865780bdafb05a564b5db28" Dec 03 02:32:56 crc kubenswrapper[4912]: I1203 02:32:56.235130 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.329258 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 02:33:00 crc kubenswrapper[4912]: E1203 02:33:00.330135 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5f2ea7f-cdc9-4573-8b16-6561c761f91f" containerName="tempest-tests-tempest-tests-runner" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.330153 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5f2ea7f-cdc9-4573-8b16-6561c761f91f" containerName="tempest-tests-tempest-tests-runner" Dec 03 02:33:00 crc kubenswrapper[4912]: E1203 02:33:00.330172 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerName="extract-utilities" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.330183 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerName="extract-utilities" Dec 03 02:33:00 crc kubenswrapper[4912]: E1203 02:33:00.330209 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerName="extract-content" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.330216 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerName="extract-content" Dec 03 02:33:00 crc kubenswrapper[4912]: E1203 02:33:00.330254 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerName="registry-server" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.330262 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerName="registry-server" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.330584 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5f2ea7f-cdc9-4573-8b16-6561c761f91f" containerName="tempest-tests-tempest-tests-runner" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.330613 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="caa09c0b-cb22-4c5c-a3a1-0269f51975ef" containerName="registry-server" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.331526 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.334201 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9xb2x" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.344150 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.459900 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5rkr\" (UniqueName: \"kubernetes.io/projected/deabdb62-423d-4c64-bbea-dd488b38ab79-kube-api-access-w5rkr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"deabdb62-423d-4c64-bbea-dd488b38ab79\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.460067 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"deabdb62-423d-4c64-bbea-dd488b38ab79\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.562239 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"deabdb62-423d-4c64-bbea-dd488b38ab79\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.562587 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5rkr\" (UniqueName: \"kubernetes.io/projected/deabdb62-423d-4c64-bbea-dd488b38ab79-kube-api-access-w5rkr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"deabdb62-423d-4c64-bbea-dd488b38ab79\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.563345 4912 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"deabdb62-423d-4c64-bbea-dd488b38ab79\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.589141 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5rkr\" (UniqueName: \"kubernetes.io/projected/deabdb62-423d-4c64-bbea-dd488b38ab79-kube-api-access-w5rkr\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"deabdb62-423d-4c64-bbea-dd488b38ab79\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.653798 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"deabdb62-423d-4c64-bbea-dd488b38ab79\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 02:33:00 crc kubenswrapper[4912]: I1203 02:33:00.956774 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 02:33:01 crc kubenswrapper[4912]: I1203 02:33:01.518189 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 02:33:02 crc kubenswrapper[4912]: I1203 02:33:02.320230 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"deabdb62-423d-4c64-bbea-dd488b38ab79","Type":"ContainerStarted","Data":"7e4a0ec476b9262342f258313327e931e844c79e7baf9c1d9b78a1b223902685"} Dec 03 02:33:02 crc kubenswrapper[4912]: I1203 02:33:02.572244 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:33:02 crc kubenswrapper[4912]: E1203 02:33:02.572786 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:33:03 crc kubenswrapper[4912]: I1203 02:33:03.335887 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"deabdb62-423d-4c64-bbea-dd488b38ab79","Type":"ContainerStarted","Data":"c18dc63d3acd4aa6e2400efb4c4c9058cf97e83e6bcadc190c8973cd51e9bafe"} Dec 03 02:33:03 crc kubenswrapper[4912]: I1203 02:33:03.356704 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.262099027 podStartE2EDuration="3.356685594s" podCreationTimestamp="2025-12-03 02:33:00 +0000 UTC" firstStartedPulling="2025-12-03 02:33:01.535810948 +0000 UTC m=+7767.177831518" lastFinishedPulling="2025-12-03 02:33:02.630397525 +0000 UTC m=+7768.272418085" observedRunningTime="2025-12-03 02:33:03.351794518 +0000 UTC m=+7768.993815098" watchObservedRunningTime="2025-12-03 02:33:03.356685594 +0000 UTC m=+7768.998706154" Dec 03 02:33:14 crc kubenswrapper[4912]: I1203 02:33:14.588561 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:33:14 crc kubenswrapper[4912]: E1203 02:33:14.590264 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:33:29 crc kubenswrapper[4912]: I1203 02:33:29.572323 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:33:29 crc kubenswrapper[4912]: E1203 02:33:29.573823 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.207950 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cj2nv/must-gather-49cx7"] Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.210065 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.212493 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-cj2nv"/"default-dockercfg-zfp2h" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.212749 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-cj2nv"/"kube-root-ca.crt" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.213290 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-cj2nv"/"openshift-service-ca.crt" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.222498 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-cj2nv/must-gather-49cx7"] Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.261590 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lv8f\" (UniqueName: \"kubernetes.io/projected/d96706ef-a068-4bab-bd26-97b7e80a8fde-kube-api-access-8lv8f\") pod \"must-gather-49cx7\" (UID: \"d96706ef-a068-4bab-bd26-97b7e80a8fde\") " pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.261675 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d96706ef-a068-4bab-bd26-97b7e80a8fde-must-gather-output\") pod \"must-gather-49cx7\" (UID: \"d96706ef-a068-4bab-bd26-97b7e80a8fde\") " pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.364034 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d96706ef-a068-4bab-bd26-97b7e80a8fde-must-gather-output\") pod \"must-gather-49cx7\" (UID: \"d96706ef-a068-4bab-bd26-97b7e80a8fde\") " pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.364457 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d96706ef-a068-4bab-bd26-97b7e80a8fde-must-gather-output\") pod \"must-gather-49cx7\" (UID: \"d96706ef-a068-4bab-bd26-97b7e80a8fde\") " pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.364701 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lv8f\" (UniqueName: \"kubernetes.io/projected/d96706ef-a068-4bab-bd26-97b7e80a8fde-kube-api-access-8lv8f\") pod \"must-gather-49cx7\" (UID: \"d96706ef-a068-4bab-bd26-97b7e80a8fde\") " pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.394547 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lv8f\" (UniqueName: \"kubernetes.io/projected/d96706ef-a068-4bab-bd26-97b7e80a8fde-kube-api-access-8lv8f\") pod \"must-gather-49cx7\" (UID: \"d96706ef-a068-4bab-bd26-97b7e80a8fde\") " pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:33:32 crc kubenswrapper[4912]: I1203 02:33:32.566846 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:33:33 crc kubenswrapper[4912]: I1203 02:33:33.291227 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-cj2nv/must-gather-49cx7"] Dec 03 02:33:33 crc kubenswrapper[4912]: I1203 02:33:33.820147 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/must-gather-49cx7" event={"ID":"d96706ef-a068-4bab-bd26-97b7e80a8fde","Type":"ContainerStarted","Data":"76b39e2a6df068296d1ad12a13e47cb2d8aeaef3333b8b33ba5eb61c032377b5"} Dec 03 02:33:38 crc kubenswrapper[4912]: I1203 02:33:38.896330 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/must-gather-49cx7" event={"ID":"d96706ef-a068-4bab-bd26-97b7e80a8fde","Type":"ContainerStarted","Data":"7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4"} Dec 03 02:33:38 crc kubenswrapper[4912]: I1203 02:33:38.897020 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/must-gather-49cx7" event={"ID":"d96706ef-a068-4bab-bd26-97b7e80a8fde","Type":"ContainerStarted","Data":"e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7"} Dec 03 02:33:38 crc kubenswrapper[4912]: I1203 02:33:38.929731 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-cj2nv/must-gather-49cx7" podStartSLOduration=2.808695637 podStartE2EDuration="6.929707392s" podCreationTimestamp="2025-12-03 02:33:32 +0000 UTC" firstStartedPulling="2025-12-03 02:33:33.294039719 +0000 UTC m=+7798.936060279" lastFinishedPulling="2025-12-03 02:33:37.415051464 +0000 UTC m=+7803.057072034" observedRunningTime="2025-12-03 02:33:38.925217088 +0000 UTC m=+7804.567237658" watchObservedRunningTime="2025-12-03 02:33:38.929707392 +0000 UTC m=+7804.571727992" Dec 03 02:33:42 crc kubenswrapper[4912]: I1203 02:33:42.752326 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cj2nv/crc-debug-gr99r"] Dec 03 02:33:42 crc kubenswrapper[4912]: I1203 02:33:42.754124 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:33:42 crc kubenswrapper[4912]: I1203 02:33:42.764895 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90dae42d-5d87-40f0-be8a-e7e0f3533b76-host\") pod \"crc-debug-gr99r\" (UID: \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\") " pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:33:42 crc kubenswrapper[4912]: I1203 02:33:42.764956 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zszl9\" (UniqueName: \"kubernetes.io/projected/90dae42d-5d87-40f0-be8a-e7e0f3533b76-kube-api-access-zszl9\") pod \"crc-debug-gr99r\" (UID: \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\") " pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:33:42 crc kubenswrapper[4912]: I1203 02:33:42.867858 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90dae42d-5d87-40f0-be8a-e7e0f3533b76-host\") pod \"crc-debug-gr99r\" (UID: \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\") " pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:33:42 crc kubenswrapper[4912]: I1203 02:33:42.867921 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zszl9\" (UniqueName: \"kubernetes.io/projected/90dae42d-5d87-40f0-be8a-e7e0f3533b76-kube-api-access-zszl9\") pod \"crc-debug-gr99r\" (UID: \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\") " pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:33:42 crc kubenswrapper[4912]: I1203 02:33:42.868263 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90dae42d-5d87-40f0-be8a-e7e0f3533b76-host\") pod \"crc-debug-gr99r\" (UID: \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\") " pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:33:42 crc kubenswrapper[4912]: I1203 02:33:42.892321 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zszl9\" (UniqueName: \"kubernetes.io/projected/90dae42d-5d87-40f0-be8a-e7e0f3533b76-kube-api-access-zszl9\") pod \"crc-debug-gr99r\" (UID: \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\") " pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:33:43 crc kubenswrapper[4912]: I1203 02:33:43.075650 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:33:43 crc kubenswrapper[4912]: I1203 02:33:43.958988 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/crc-debug-gr99r" event={"ID":"90dae42d-5d87-40f0-be8a-e7e0f3533b76","Type":"ContainerStarted","Data":"1a6b60e1afe639b5a87fd006850bace92ac2ab2c076a203691a1e06c63428ca1"} Dec 03 02:33:44 crc kubenswrapper[4912]: I1203 02:33:44.580590 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:33:44 crc kubenswrapper[4912]: E1203 02:33:44.581142 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:33:56 crc kubenswrapper[4912]: I1203 02:33:56.097238 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/crc-debug-gr99r" event={"ID":"90dae42d-5d87-40f0-be8a-e7e0f3533b76","Type":"ContainerStarted","Data":"24b107239d48a46b8f33ad1cbfe16feae80f1b2d7f54a7fb7c1b9d7701e94bbd"} Dec 03 02:33:56 crc kubenswrapper[4912]: I1203 02:33:56.122811 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-cj2nv/crc-debug-gr99r" podStartSLOduration=2.092130579 podStartE2EDuration="14.12279339s" podCreationTimestamp="2025-12-03 02:33:42 +0000 UTC" firstStartedPulling="2025-12-03 02:33:43.128919253 +0000 UTC m=+7808.770939813" lastFinishedPulling="2025-12-03 02:33:55.159582054 +0000 UTC m=+7820.801602624" observedRunningTime="2025-12-03 02:33:56.120212618 +0000 UTC m=+7821.762233188" watchObservedRunningTime="2025-12-03 02:33:56.12279339 +0000 UTC m=+7821.764813950" Dec 03 02:33:57 crc kubenswrapper[4912]: I1203 02:33:57.571778 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:33:57 crc kubenswrapper[4912]: E1203 02:33:57.572693 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:34:09 crc kubenswrapper[4912]: I1203 02:34:09.571708 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:34:09 crc kubenswrapper[4912]: E1203 02:34:09.572517 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:34:21 crc kubenswrapper[4912]: I1203 02:34:21.571735 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:34:21 crc kubenswrapper[4912]: E1203 02:34:21.573242 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:34:35 crc kubenswrapper[4912]: I1203 02:34:35.571984 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:34:35 crc kubenswrapper[4912]: E1203 02:34:35.573310 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:34:43 crc kubenswrapper[4912]: I1203 02:34:43.735581 4912 generic.go:334] "Generic (PLEG): container finished" podID="90dae42d-5d87-40f0-be8a-e7e0f3533b76" containerID="24b107239d48a46b8f33ad1cbfe16feae80f1b2d7f54a7fb7c1b9d7701e94bbd" exitCode=0 Dec 03 02:34:43 crc kubenswrapper[4912]: I1203 02:34:43.735672 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/crc-debug-gr99r" event={"ID":"90dae42d-5d87-40f0-be8a-e7e0f3533b76","Type":"ContainerDied","Data":"24b107239d48a46b8f33ad1cbfe16feae80f1b2d7f54a7fb7c1b9d7701e94bbd"} Dec 03 02:34:44 crc kubenswrapper[4912]: I1203 02:34:44.860610 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:34:44 crc kubenswrapper[4912]: I1203 02:34:44.895024 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cj2nv/crc-debug-gr99r"] Dec 03 02:34:44 crc kubenswrapper[4912]: I1203 02:34:44.909118 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cj2nv/crc-debug-gr99r"] Dec 03 02:34:44 crc kubenswrapper[4912]: I1203 02:34:44.951708 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90dae42d-5d87-40f0-be8a-e7e0f3533b76-host\") pod \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\" (UID: \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\") " Dec 03 02:34:44 crc kubenswrapper[4912]: I1203 02:34:44.951803 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/90dae42d-5d87-40f0-be8a-e7e0f3533b76-host" (OuterVolumeSpecName: "host") pod "90dae42d-5d87-40f0-be8a-e7e0f3533b76" (UID: "90dae42d-5d87-40f0-be8a-e7e0f3533b76"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 02:34:44 crc kubenswrapper[4912]: I1203 02:34:44.951942 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zszl9\" (UniqueName: \"kubernetes.io/projected/90dae42d-5d87-40f0-be8a-e7e0f3533b76-kube-api-access-zszl9\") pod \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\" (UID: \"90dae42d-5d87-40f0-be8a-e7e0f3533b76\") " Dec 03 02:34:44 crc kubenswrapper[4912]: I1203 02:34:44.952508 4912 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90dae42d-5d87-40f0-be8a-e7e0f3533b76-host\") on node \"crc\" DevicePath \"\"" Dec 03 02:34:44 crc kubenswrapper[4912]: I1203 02:34:44.973491 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90dae42d-5d87-40f0-be8a-e7e0f3533b76-kube-api-access-zszl9" (OuterVolumeSpecName: "kube-api-access-zszl9") pod "90dae42d-5d87-40f0-be8a-e7e0f3533b76" (UID: "90dae42d-5d87-40f0-be8a-e7e0f3533b76"). InnerVolumeSpecName "kube-api-access-zszl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:34:45 crc kubenswrapper[4912]: I1203 02:34:45.054046 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zszl9\" (UniqueName: \"kubernetes.io/projected/90dae42d-5d87-40f0-be8a-e7e0f3533b76-kube-api-access-zszl9\") on node \"crc\" DevicePath \"\"" Dec 03 02:34:45 crc kubenswrapper[4912]: I1203 02:34:45.760853 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a6b60e1afe639b5a87fd006850bace92ac2ab2c076a203691a1e06c63428ca1" Dec 03 02:34:45 crc kubenswrapper[4912]: I1203 02:34:45.761352 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-gr99r" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.060462 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cj2nv/crc-debug-8llwd"] Dec 03 02:34:46 crc kubenswrapper[4912]: E1203 02:34:46.060850 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90dae42d-5d87-40f0-be8a-e7e0f3533b76" containerName="container-00" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.060862 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="90dae42d-5d87-40f0-be8a-e7e0f3533b76" containerName="container-00" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.061103 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="90dae42d-5d87-40f0-be8a-e7e0f3533b76" containerName="container-00" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.061857 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.178062 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bb57\" (UniqueName: \"kubernetes.io/projected/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-kube-api-access-9bb57\") pod \"crc-debug-8llwd\" (UID: \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\") " pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.178834 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-host\") pod \"crc-debug-8llwd\" (UID: \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\") " pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.280741 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-host\") pod \"crc-debug-8llwd\" (UID: \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\") " pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.280896 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-host\") pod \"crc-debug-8llwd\" (UID: \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\") " pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.280904 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bb57\" (UniqueName: \"kubernetes.io/projected/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-kube-api-access-9bb57\") pod \"crc-debug-8llwd\" (UID: \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\") " pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.301108 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bb57\" (UniqueName: \"kubernetes.io/projected/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-kube-api-access-9bb57\") pod \"crc-debug-8llwd\" (UID: \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\") " pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.382942 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.588117 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90dae42d-5d87-40f0-be8a-e7e0f3533b76" path="/var/lib/kubelet/pods/90dae42d-5d87-40f0-be8a-e7e0f3533b76/volumes" Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.772155 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/crc-debug-8llwd" event={"ID":"f6f17cea-3361-4e71-9385-5ccd73fe7cb6","Type":"ContainerStarted","Data":"70691aaef0b9b021629063f1fd4e166b881957e9a1f292acf3b029cef3c17031"} Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.772213 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/crc-debug-8llwd" event={"ID":"f6f17cea-3361-4e71-9385-5ccd73fe7cb6","Type":"ContainerStarted","Data":"60dec3ce074f0494cbd3f9a6dbecb40a93d3077fe02c7f7de0fa3619c80342be"} Dec 03 02:34:46 crc kubenswrapper[4912]: I1203 02:34:46.805577 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-cj2nv/crc-debug-8llwd" podStartSLOduration=0.805552583 podStartE2EDuration="805.552583ms" podCreationTimestamp="2025-12-03 02:34:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 02:34:46.785792475 +0000 UTC m=+7872.427813065" watchObservedRunningTime="2025-12-03 02:34:46.805552583 +0000 UTC m=+7872.447573163" Dec 03 02:34:47 crc kubenswrapper[4912]: I1203 02:34:47.782011 4912 generic.go:334] "Generic (PLEG): container finished" podID="f6f17cea-3361-4e71-9385-5ccd73fe7cb6" containerID="70691aaef0b9b021629063f1fd4e166b881957e9a1f292acf3b029cef3c17031" exitCode=0 Dec 03 02:34:47 crc kubenswrapper[4912]: I1203 02:34:47.782093 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/crc-debug-8llwd" event={"ID":"f6f17cea-3361-4e71-9385-5ccd73fe7cb6","Type":"ContainerDied","Data":"70691aaef0b9b021629063f1fd4e166b881957e9a1f292acf3b029cef3c17031"} Dec 03 02:34:48 crc kubenswrapper[4912]: I1203 02:34:48.903564 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.038049 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bb57\" (UniqueName: \"kubernetes.io/projected/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-kube-api-access-9bb57\") pod \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\" (UID: \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\") " Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.038088 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-host\") pod \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\" (UID: \"f6f17cea-3361-4e71-9385-5ccd73fe7cb6\") " Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.038455 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-host" (OuterVolumeSpecName: "host") pod "f6f17cea-3361-4e71-9385-5ccd73fe7cb6" (UID: "f6f17cea-3361-4e71-9385-5ccd73fe7cb6"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.038746 4912 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-host\") on node \"crc\" DevicePath \"\"" Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.048312 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-kube-api-access-9bb57" (OuterVolumeSpecName: "kube-api-access-9bb57") pod "f6f17cea-3361-4e71-9385-5ccd73fe7cb6" (UID: "f6f17cea-3361-4e71-9385-5ccd73fe7cb6"). InnerVolumeSpecName "kube-api-access-9bb57". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.140307 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bb57\" (UniqueName: \"kubernetes.io/projected/f6f17cea-3361-4e71-9385-5ccd73fe7cb6-kube-api-access-9bb57\") on node \"crc\" DevicePath \"\"" Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.430663 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cj2nv/crc-debug-8llwd"] Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.440356 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cj2nv/crc-debug-8llwd"] Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.814158 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60dec3ce074f0494cbd3f9a6dbecb40a93d3077fe02c7f7de0fa3619c80342be" Dec 03 02:34:49 crc kubenswrapper[4912]: I1203 02:34:49.814258 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-8llwd" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.571787 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:34:50 crc kubenswrapper[4912]: E1203 02:34:50.572246 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.589261 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6f17cea-3361-4e71-9385-5ccd73fe7cb6" path="/var/lib/kubelet/pods/f6f17cea-3361-4e71-9385-5ccd73fe7cb6/volumes" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.613788 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cj2nv/crc-debug-pqd77"] Dec 03 02:34:50 crc kubenswrapper[4912]: E1203 02:34:50.614540 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f17cea-3361-4e71-9385-5ccd73fe7cb6" containerName="container-00" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.614732 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f17cea-3361-4e71-9385-5ccd73fe7cb6" containerName="container-00" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.615032 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f17cea-3361-4e71-9385-5ccd73fe7cb6" containerName="container-00" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.615866 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.777795 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxd8k\" (UniqueName: \"kubernetes.io/projected/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-kube-api-access-wxd8k\") pod \"crc-debug-pqd77\" (UID: \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\") " pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.778106 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-host\") pod \"crc-debug-pqd77\" (UID: \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\") " pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.880354 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxd8k\" (UniqueName: \"kubernetes.io/projected/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-kube-api-access-wxd8k\") pod \"crc-debug-pqd77\" (UID: \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\") " pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.880768 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-host\") pod \"crc-debug-pqd77\" (UID: \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\") " pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.880932 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-host\") pod \"crc-debug-pqd77\" (UID: \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\") " pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.902942 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxd8k\" (UniqueName: \"kubernetes.io/projected/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-kube-api-access-wxd8k\") pod \"crc-debug-pqd77\" (UID: \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\") " pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:50 crc kubenswrapper[4912]: I1203 02:34:50.935222 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:50 crc kubenswrapper[4912]: W1203 02:34:50.974073 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa8b1cc4_f280_46f9_b8e1_17f465e00f61.slice/crio-d2ee43a6987a6548855bbcb61983476f50ab52adff52f78279dae2fe435e13e9 WatchSource:0}: Error finding container d2ee43a6987a6548855bbcb61983476f50ab52adff52f78279dae2fe435e13e9: Status 404 returned error can't find the container with id d2ee43a6987a6548855bbcb61983476f50ab52adff52f78279dae2fe435e13e9 Dec 03 02:34:51 crc kubenswrapper[4912]: I1203 02:34:51.842960 4912 generic.go:334] "Generic (PLEG): container finished" podID="fa8b1cc4-f280-46f9-b8e1-17f465e00f61" containerID="3c90154e2fbf48ef9ca8d0c4d7263ac8e34cfc9581cc7dde77cf1e35e3581a91" exitCode=0 Dec 03 02:34:51 crc kubenswrapper[4912]: I1203 02:34:51.843031 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/crc-debug-pqd77" event={"ID":"fa8b1cc4-f280-46f9-b8e1-17f465e00f61","Type":"ContainerDied","Data":"3c90154e2fbf48ef9ca8d0c4d7263ac8e34cfc9581cc7dde77cf1e35e3581a91"} Dec 03 02:34:51 crc kubenswrapper[4912]: I1203 02:34:51.843364 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/crc-debug-pqd77" event={"ID":"fa8b1cc4-f280-46f9-b8e1-17f465e00f61","Type":"ContainerStarted","Data":"d2ee43a6987a6548855bbcb61983476f50ab52adff52f78279dae2fe435e13e9"} Dec 03 02:34:51 crc kubenswrapper[4912]: I1203 02:34:51.896228 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cj2nv/crc-debug-pqd77"] Dec 03 02:34:51 crc kubenswrapper[4912]: I1203 02:34:51.915488 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cj2nv/crc-debug-pqd77"] Dec 03 02:34:53 crc kubenswrapper[4912]: I1203 02:34:53.000300 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:53 crc kubenswrapper[4912]: I1203 02:34:53.133309 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxd8k\" (UniqueName: \"kubernetes.io/projected/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-kube-api-access-wxd8k\") pod \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\" (UID: \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\") " Dec 03 02:34:53 crc kubenswrapper[4912]: I1203 02:34:53.133850 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-host\") pod \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\" (UID: \"fa8b1cc4-f280-46f9-b8e1-17f465e00f61\") " Dec 03 02:34:53 crc kubenswrapper[4912]: I1203 02:34:53.133982 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-host" (OuterVolumeSpecName: "host") pod "fa8b1cc4-f280-46f9-b8e1-17f465e00f61" (UID: "fa8b1cc4-f280-46f9-b8e1-17f465e00f61"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 02:34:53 crc kubenswrapper[4912]: I1203 02:34:53.134587 4912 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-host\") on node \"crc\" DevicePath \"\"" Dec 03 02:34:53 crc kubenswrapper[4912]: I1203 02:34:53.143095 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-kube-api-access-wxd8k" (OuterVolumeSpecName: "kube-api-access-wxd8k") pod "fa8b1cc4-f280-46f9-b8e1-17f465e00f61" (UID: "fa8b1cc4-f280-46f9-b8e1-17f465e00f61"). InnerVolumeSpecName "kube-api-access-wxd8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:34:53 crc kubenswrapper[4912]: I1203 02:34:53.237183 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxd8k\" (UniqueName: \"kubernetes.io/projected/fa8b1cc4-f280-46f9-b8e1-17f465e00f61-kube-api-access-wxd8k\") on node \"crc\" DevicePath \"\"" Dec 03 02:34:53 crc kubenswrapper[4912]: I1203 02:34:53.866474 4912 scope.go:117] "RemoveContainer" containerID="3c90154e2fbf48ef9ca8d0c4d7263ac8e34cfc9581cc7dde77cf1e35e3581a91" Dec 03 02:34:53 crc kubenswrapper[4912]: I1203 02:34:53.866549 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/crc-debug-pqd77" Dec 03 02:34:54 crc kubenswrapper[4912]: I1203 02:34:54.612989 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa8b1cc4-f280-46f9-b8e1-17f465e00f61" path="/var/lib/kubelet/pods/fa8b1cc4-f280-46f9-b8e1-17f465e00f61/volumes" Dec 03 02:35:03 crc kubenswrapper[4912]: I1203 02:35:03.572022 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:35:03 crc kubenswrapper[4912]: E1203 02:35:03.572867 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:35:15 crc kubenswrapper[4912]: I1203 02:35:15.573923 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:35:15 crc kubenswrapper[4912]: E1203 02:35:15.574633 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:35:28 crc kubenswrapper[4912]: I1203 02:35:28.571774 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:35:28 crc kubenswrapper[4912]: E1203 02:35:28.572371 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:35:43 crc kubenswrapper[4912]: I1203 02:35:43.572687 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:35:43 crc kubenswrapper[4912]: E1203 02:35:43.574030 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.104717 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_41945ca7-6d5e-4875-b770-801813413434/aodh-api/0.log" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.301050 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_41945ca7-6d5e-4875-b770-801813413434/aodh-notifier/0.log" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.301129 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_41945ca7-6d5e-4875-b770-801813413434/aodh-evaluator/0.log" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.327135 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_41945ca7-6d5e-4875-b770-801813413434/aodh-listener/0.log" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.558020 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-75f549cc78-m5rcp_21e2b86e-eeab-4af6-b68d-8f67abda11e2/barbican-api/0.log" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.575548 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-75f549cc78-m5rcp_21e2b86e-eeab-4af6-b68d-8f67abda11e2/barbican-api-log/0.log" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.617743 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6f4c446894-wlqxw_3269ec0e-5405-4a8d-b7ea-e07c5a7a7212/barbican-keystone-listener/0.log" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.764398 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-79459d7cd7-qdq5s_c6d0b29e-b4cd-4f78-b9f7-200691c64f93/barbican-worker/0.log" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.850533 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-79459d7cd7-qdq5s_c6d0b29e-b4cd-4f78-b9f7-200691c64f93/barbican-worker-log/0.log" Dec 03 02:35:46 crc kubenswrapper[4912]: I1203 02:35:46.877345 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6f4c446894-wlqxw_3269ec0e-5405-4a8d-b7ea-e07c5a7a7212/barbican-keystone-listener-log/0.log" Dec 03 02:35:47 crc kubenswrapper[4912]: I1203 02:35:47.051196 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr_ac32e469-f344-4ee1-8702-bfa8b124ac35/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:47 crc kubenswrapper[4912]: I1203 02:35:47.126151 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bbc8b9dd-35e3-4d6b-8df6-e44368688196/ceilometer-central-agent/0.log" Dec 03 02:35:47 crc kubenswrapper[4912]: I1203 02:35:47.272476 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bbc8b9dd-35e3-4d6b-8df6-e44368688196/ceilometer-notification-agent/0.log" Dec 03 02:35:47 crc kubenswrapper[4912]: I1203 02:35:47.315553 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bbc8b9dd-35e3-4d6b-8df6-e44368688196/proxy-httpd/0.log" Dec 03 02:35:47 crc kubenswrapper[4912]: I1203 02:35:47.406397 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bbc8b9dd-35e3-4d6b-8df6-e44368688196/sg-core/0.log" Dec 03 02:35:47 crc kubenswrapper[4912]: I1203 02:35:47.521180 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv_b529c480-6fb1-4a0b-a915-1aca2f6e85bd/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:47 crc kubenswrapper[4912]: I1203 02:35:47.658757 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg_d241940f-a3d4-4c91-93c7-2149a6e25ae2/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:47 crc kubenswrapper[4912]: I1203 02:35:47.834096 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_64841d14-78b5-4737-90ec-b812f2994397/cinder-api-log/0.log" Dec 03 02:35:47 crc kubenswrapper[4912]: I1203 02:35:47.868477 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_64841d14-78b5-4737-90ec-b812f2994397/cinder-api/0.log" Dec 03 02:35:48 crc kubenswrapper[4912]: I1203 02:35:48.114437 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7/probe/0.log" Dec 03 02:35:48 crc kubenswrapper[4912]: I1203 02:35:48.117596 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7/cinder-backup/0.log" Dec 03 02:35:48 crc kubenswrapper[4912]: I1203 02:35:48.460725 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0/cinder-scheduler/0.log" Dec 03 02:35:48 crc kubenswrapper[4912]: I1203 02:35:48.551930 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0/probe/0.log" Dec 03 02:35:48 crc kubenswrapper[4912]: I1203 02:35:48.595159 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3e1c22b1-b578-4957-98b5-b981f1f7dee2/cinder-volume/0.log" Dec 03 02:35:48 crc kubenswrapper[4912]: I1203 02:35:48.787126 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3e1c22b1-b578-4957-98b5-b981f1f7dee2/probe/0.log" Dec 03 02:35:48 crc kubenswrapper[4912]: I1203 02:35:48.795798 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv_50268f51-4c6d-4837-8806-3f741ee7b48c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:49 crc kubenswrapper[4912]: I1203 02:35:49.041186 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-p2gz6_51b6bb89-372c-4193-ab71-6317807c1820/init/0.log" Dec 03 02:35:49 crc kubenswrapper[4912]: I1203 02:35:49.127367 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-78l9p_12198d9f-b54f-428d-9a81-7d6d2d44f695/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:49 crc kubenswrapper[4912]: I1203 02:35:49.261018 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-p2gz6_51b6bb89-372c-4193-ab71-6317807c1820/init/0.log" Dec 03 02:35:49 crc kubenswrapper[4912]: I1203 02:35:49.338629 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-p2gz6_51b6bb89-372c-4193-ab71-6317807c1820/dnsmasq-dns/0.log" Dec 03 02:35:49 crc kubenswrapper[4912]: I1203 02:35:49.340939 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bf551edf-1401-42c2-8d75-6dec09a149c6/glance-httpd/0.log" Dec 03 02:35:49 crc kubenswrapper[4912]: I1203 02:35:49.498122 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bf551edf-1401-42c2-8d75-6dec09a149c6/glance-log/0.log" Dec 03 02:35:49 crc kubenswrapper[4912]: I1203 02:35:49.576025 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17260561-5216-45fa-b547-2cbe178094d1/glance-httpd/0.log" Dec 03 02:35:49 crc kubenswrapper[4912]: I1203 02:35:49.631804 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17260561-5216-45fa-b547-2cbe178094d1/glance-log/0.log" Dec 03 02:35:50 crc kubenswrapper[4912]: I1203 02:35:50.068991 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-678985f55c-wftgz_e31eb41a-2842-45fe-ba2d-6a84a048723f/heat-engine/0.log" Dec 03 02:35:50 crc kubenswrapper[4912]: I1203 02:35:50.531895 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d4dbbf8d-5vgmn_3d92b6a2-0079-4d39-9c11-9a31c22429f4/horizon/0.log" Dec 03 02:35:50 crc kubenswrapper[4912]: I1203 02:35:50.569390 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-ssn98_d2d7b572-6166-47b3-9346-17f881d869a7/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:50 crc kubenswrapper[4912]: I1203 02:35:50.869294 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-jckxr_90fce4e2-3ec4-469d-a50a-8c8df048c2ab/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:50 crc kubenswrapper[4912]: I1203 02:35:50.905015 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-56b4585b94-qjwmb_23ac5ef2-d307-4073-982f-afe777413b41/heat-api/0.log" Dec 03 02:35:50 crc kubenswrapper[4912]: I1203 02:35:50.938872 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5cdb6985d5-xgpkb_b0e666e2-5db8-409a-9a47-0da8fe895f75/heat-cfnapi/0.log" Dec 03 02:35:50 crc kubenswrapper[4912]: I1203 02:35:50.984159 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d4dbbf8d-5vgmn_3d92b6a2-0079-4d39-9c11-9a31c22429f4/horizon-log/0.log" Dec 03 02:35:51 crc kubenswrapper[4912]: I1203 02:35:51.136760 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412061-jvbbz_65a4a770-16d8-4ea3-ac4a-9810ed0c2477/keystone-cron/0.log" Dec 03 02:35:51 crc kubenswrapper[4912]: I1203 02:35:51.181884 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412121-dkhrq_bb6b4855-3d83-4341-942e-de0bc1027bc4/keystone-cron/0.log" Dec 03 02:35:51 crc kubenswrapper[4912]: I1203 02:35:51.419109 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba/kube-state-metrics/0.log" Dec 03 02:35:51 crc kubenswrapper[4912]: I1203 02:35:51.603368 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-q9btl_911b9678-7127-4ae0-ba4b-c059dc13796f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:51 crc kubenswrapper[4912]: I1203 02:35:51.633910 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-46v4t_77e171a1-53b6-4a5d-94a2-2d9b42e1e71f/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:51 crc kubenswrapper[4912]: I1203 02:35:51.742136 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-d9977975b-bq8m5_a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b/keystone-api/0.log" Dec 03 02:35:51 crc kubenswrapper[4912]: I1203 02:35:51.870304 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_eedcff7d-7bb2-4c30-8166-e7a2f22ac60e/manila-api-log/0.log" Dec 03 02:35:51 crc kubenswrapper[4912]: I1203 02:35:51.915128 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_eedcff7d-7bb2-4c30-8166-e7a2f22ac60e/manila-api/0.log" Dec 03 02:35:51 crc kubenswrapper[4912]: I1203 02:35:51.971022 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_dc97be40-fb7b-4af2-af77-45f0f360c8cb/probe/0.log" Dec 03 02:35:52 crc kubenswrapper[4912]: I1203 02:35:52.026934 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_dc97be40-fb7b-4af2-af77-45f0f360c8cb/manila-scheduler/0.log" Dec 03 02:35:52 crc kubenswrapper[4912]: I1203 02:35:52.124959 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b2a55784-64e0-4397-a975-c0b069cbd08f/manila-share/0.log" Dec 03 02:35:52 crc kubenswrapper[4912]: I1203 02:35:52.169558 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b2a55784-64e0-4397-a975-c0b069cbd08f/probe/0.log" Dec 03 02:35:52 crc kubenswrapper[4912]: I1203 02:35:52.359918 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_610be762-7f1b-4435-8764-eeac647f901c/mysqld-exporter/0.log" Dec 03 02:35:52 crc kubenswrapper[4912]: I1203 02:35:52.878201 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59fd659dd5-vmcdc_6b4100ae-eb0e-402b-b442-d2d313e2bff4/neutron-api/0.log" Dec 03 02:35:52 crc kubenswrapper[4912]: I1203 02:35:52.933773 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747_748f1896-9a69-406c-bc59-0662de1da921/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:52 crc kubenswrapper[4912]: I1203 02:35:52.936292 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59fd659dd5-vmcdc_6b4100ae-eb0e-402b-b442-d2d313e2bff4/neutron-httpd/0.log" Dec 03 02:35:53 crc kubenswrapper[4912]: I1203 02:35:53.463553 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2d46d472-d813-4f6d-b58a-8d671244960a/nova-cell0-conductor-conductor/0.log" Dec 03 02:35:53 crc kubenswrapper[4912]: I1203 02:35:53.779122 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_8116ee8d-aa67-4a8f-a765-b846f397f7d4/nova-cell1-conductor-conductor/0.log" Dec 03 02:35:53 crc kubenswrapper[4912]: I1203 02:35:53.819210 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23c2e51b-07f1-4b86-bc02-29969db68f38/nova-api-log/0.log" Dec 03 02:35:54 crc kubenswrapper[4912]: I1203 02:35:54.031838 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c73f5ab7-bc17-41fa-80af-55fbb2ebf0de/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 02:35:54 crc kubenswrapper[4912]: I1203 02:35:54.101955 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8_d65dbc95-9df7-49ec-9d0a-b883b70b4e3e/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:54 crc kubenswrapper[4912]: I1203 02:35:54.332996 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ba93637e-3299-4ead-926f-0e2f84c31df9/nova-metadata-log/0.log" Dec 03 02:35:54 crc kubenswrapper[4912]: I1203 02:35:54.559829 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23c2e51b-07f1-4b86-bc02-29969db68f38/nova-api-api/0.log" Dec 03 02:35:54 crc kubenswrapper[4912]: I1203 02:35:54.734677 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e6cbac08-d327-4238-8335-a69cfb3f71b4/nova-scheduler-scheduler/0.log" Dec 03 02:35:54 crc kubenswrapper[4912]: I1203 02:35:54.771147 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e835f70a-04f2-43d3-ba92-6a6d20216a7d/mysql-bootstrap/0.log" Dec 03 02:35:54 crc kubenswrapper[4912]: I1203 02:35:54.955401 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e835f70a-04f2-43d3-ba92-6a6d20216a7d/mysql-bootstrap/0.log" Dec 03 02:35:55 crc kubenswrapper[4912]: I1203 02:35:55.036119 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e835f70a-04f2-43d3-ba92-6a6d20216a7d/galera/0.log" Dec 03 02:35:55 crc kubenswrapper[4912]: I1203 02:35:55.188525 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9a1a55bf-bc0f-4004-b540-ead98e30d25a/mysql-bootstrap/0.log" Dec 03 02:35:55 crc kubenswrapper[4912]: I1203 02:35:55.344788 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9a1a55bf-bc0f-4004-b540-ead98e30d25a/mysql-bootstrap/0.log" Dec 03 02:35:55 crc kubenswrapper[4912]: I1203 02:35:55.399790 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9a1a55bf-bc0f-4004-b540-ead98e30d25a/galera/0.log" Dec 03 02:35:55 crc kubenswrapper[4912]: I1203 02:35:55.560598 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b1cf8990-bf85-4de8-8121-73c43a7eccd9/openstackclient/0.log" Dec 03 02:35:55 crc kubenswrapper[4912]: I1203 02:35:55.646593 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-gbbh7_dd4efef6-5ce3-444a-a464-1f18f7cc2db4/ovn-controller/0.log" Dec 03 02:35:55 crc kubenswrapper[4912]: I1203 02:35:55.791798 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-p7c9n_290c5de7-2853-45ef-9b03-99a0d301d51f/openstack-network-exporter/0.log" Dec 03 02:35:55 crc kubenswrapper[4912]: I1203 02:35:55.998327 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6lm7j_2c613cf4-0585-466f-8db2-4c4f4b4765f0/ovsdb-server-init/0.log" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.203839 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6lm7j_2c613cf4-0585-466f-8db2-4c4f4b4765f0/ovs-vswitchd/0.log" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.235713 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6lm7j_2c613cf4-0585-466f-8db2-4c4f4b4765f0/ovsdb-server-init/0.log" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.277715 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6lm7j_2c613cf4-0585-466f-8db2-4c4f4b4765f0/ovsdb-server/0.log" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.452573 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-49qvj_04c823be-fed9-4ffd-b41b-30115e049634/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.571145 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:35:56 crc kubenswrapper[4912]: E1203 02:35:56.571441 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.621157 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_197a0bff-4ebc-4581-bc0f-77d459500339/ovn-northd/0.log" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.654552 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_197a0bff-4ebc-4581-bc0f-77d459500339/openstack-network-exporter/0.log" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.808492 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026/openstack-network-exporter/0.log" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.840207 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026/ovsdbserver-nb/0.log" Dec 03 02:35:56 crc kubenswrapper[4912]: I1203 02:35:56.985015 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d0dca179-6a1f-41ad-bad5-1d193e2583c8/openstack-network-exporter/0.log" Dec 03 02:35:57 crc kubenswrapper[4912]: I1203 02:35:57.019297 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d0dca179-6a1f-41ad-bad5-1d193e2583c8/ovsdbserver-sb/0.log" Dec 03 02:35:57 crc kubenswrapper[4912]: I1203 02:35:57.389636 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ba93637e-3299-4ead-926f-0e2f84c31df9/nova-metadata-metadata/0.log" Dec 03 02:35:57 crc kubenswrapper[4912]: I1203 02:35:57.486811 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69d59dff98-dh972_7eb76b80-5bef-4a43-a3d0-098c3029fcb2/placement-api/0.log" Dec 03 02:35:57 crc kubenswrapper[4912]: I1203 02:35:57.514031 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69d59dff98-dh972_7eb76b80-5bef-4a43-a3d0-098c3029fcb2/placement-log/0.log" Dec 03 02:35:57 crc kubenswrapper[4912]: I1203 02:35:57.582129 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/init-config-reloader/0.log" Dec 03 02:35:57 crc kubenswrapper[4912]: I1203 02:35:57.854206 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/init-config-reloader/0.log" Dec 03 02:35:57 crc kubenswrapper[4912]: I1203 02:35:57.856139 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/config-reloader/0.log" Dec 03 02:35:57 crc kubenswrapper[4912]: I1203 02:35:57.875692 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/thanos-sidecar/0.log" Dec 03 02:35:57 crc kubenswrapper[4912]: I1203 02:35:57.895308 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/prometheus/0.log" Dec 03 02:35:58 crc kubenswrapper[4912]: I1203 02:35:58.061760 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1435b2ef-93c1-41ac-a2bd-f9af839b7894/setup-container/0.log" Dec 03 02:35:58 crc kubenswrapper[4912]: I1203 02:35:58.284300 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1435b2ef-93c1-41ac-a2bd-f9af839b7894/setup-container/0.log" Dec 03 02:35:58 crc kubenswrapper[4912]: I1203 02:35:58.318281 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb7b3762-41c4-4935-8882-526576167e0a/setup-container/0.log" Dec 03 02:35:58 crc kubenswrapper[4912]: I1203 02:35:58.358514 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1435b2ef-93c1-41ac-a2bd-f9af839b7894/rabbitmq/0.log" Dec 03 02:35:58 crc kubenswrapper[4912]: I1203 02:35:58.624149 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb7b3762-41c4-4935-8882-526576167e0a/setup-container/0.log" Dec 03 02:35:58 crc kubenswrapper[4912]: I1203 02:35:58.632323 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb7b3762-41c4-4935-8882-526576167e0a/rabbitmq/0.log" Dec 03 02:35:58 crc kubenswrapper[4912]: I1203 02:35:58.661637 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh_4d2c5206-f0c2-4df9-9a99-345471f00226/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:58 crc kubenswrapper[4912]: I1203 02:35:58.877604 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz_0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:58 crc kubenswrapper[4912]: I1203 02:35:58.878777 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-6kzw8_d31b18ef-e64f-47ac-9aad-401795f9e3f0/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.033209 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-nfq59_58be2429-a01c-42d9-b773-da6cae52f9b3/ssh-known-hosts-edpm-deployment/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.311740 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7ddd66b7bc-vwpcg_540979f8-98e6-40f8-9f82-1e7b924d984e/proxy-server/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.405658 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6wv8k_1556422b-130c-4bf1-8010-ee7c6ea5f662/swift-ring-rebalance/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.480734 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7ddd66b7bc-vwpcg_540979f8-98e6-40f8-9f82-1e7b924d984e/proxy-httpd/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.620922 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/account-reaper/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.624473 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/account-auditor/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.734841 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/account-replicator/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.811254 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/account-server/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.844015 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/container-auditor/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.863984 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/container-replicator/0.log" Dec 03 02:35:59 crc kubenswrapper[4912]: I1203 02:35:59.942225 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/container-server/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.055637 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/container-updater/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.084156 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-auditor/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.129517 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-expirer/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.188719 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-replicator/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.272931 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-server/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.329472 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/rsync/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.331733 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-updater/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.449002 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/swift-recon-cron/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.574505 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng_d6d6c301-1060-4d03-8e8e-6d776c912cc2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.719970 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm_e34cc78e-e946-4d45-a3da-962e308ad972/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:36:00 crc kubenswrapper[4912]: I1203 02:36:00.940384 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_deabdb62-423d-4c64-bbea-dd488b38ab79/test-operator-logs-container/0.log" Dec 03 02:36:01 crc kubenswrapper[4912]: I1203 02:36:01.115167 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-c75zj_078c0c55-2cdc-4164-af87-c60aced0babf/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:36:01 crc kubenswrapper[4912]: I1203 02:36:01.535794 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d5f2ea7f-cdc9-4573-8b16-6561c761f91f/tempest-tests-tempest-tests-runner/0.log" Dec 03 02:36:09 crc kubenswrapper[4912]: I1203 02:36:09.571823 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:36:09 crc kubenswrapper[4912]: E1203 02:36:09.572651 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:36:10 crc kubenswrapper[4912]: I1203 02:36:10.842786 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j9g9c"] Dec 03 02:36:10 crc kubenswrapper[4912]: E1203 02:36:10.843849 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8b1cc4-f280-46f9-b8e1-17f465e00f61" containerName="container-00" Dec 03 02:36:10 crc kubenswrapper[4912]: I1203 02:36:10.843862 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8b1cc4-f280-46f9-b8e1-17f465e00f61" containerName="container-00" Dec 03 02:36:10 crc kubenswrapper[4912]: I1203 02:36:10.844089 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa8b1cc4-f280-46f9-b8e1-17f465e00f61" containerName="container-00" Dec 03 02:36:10 crc kubenswrapper[4912]: I1203 02:36:10.845917 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:10 crc kubenswrapper[4912]: I1203 02:36:10.852390 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9g9c"] Dec 03 02:36:10 crc kubenswrapper[4912]: I1203 02:36:10.954205 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgjpx\" (UniqueName: \"kubernetes.io/projected/2f1fd41c-b9f1-4008-8e17-3043627bbe72-kube-api-access-pgjpx\") pod \"redhat-marketplace-j9g9c\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:10 crc kubenswrapper[4912]: I1203 02:36:10.954668 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-utilities\") pod \"redhat-marketplace-j9g9c\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:10 crc kubenswrapper[4912]: I1203 02:36:10.954714 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-catalog-content\") pod \"redhat-marketplace-j9g9c\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:11 crc kubenswrapper[4912]: I1203 02:36:11.056204 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-utilities\") pod \"redhat-marketplace-j9g9c\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:11 crc kubenswrapper[4912]: I1203 02:36:11.056255 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-catalog-content\") pod \"redhat-marketplace-j9g9c\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:11 crc kubenswrapper[4912]: I1203 02:36:11.056302 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgjpx\" (UniqueName: \"kubernetes.io/projected/2f1fd41c-b9f1-4008-8e17-3043627bbe72-kube-api-access-pgjpx\") pod \"redhat-marketplace-j9g9c\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:11 crc kubenswrapper[4912]: I1203 02:36:11.057462 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-utilities\") pod \"redhat-marketplace-j9g9c\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:11 crc kubenswrapper[4912]: I1203 02:36:11.057947 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-catalog-content\") pod \"redhat-marketplace-j9g9c\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:11 crc kubenswrapper[4912]: I1203 02:36:11.086473 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgjpx\" (UniqueName: \"kubernetes.io/projected/2f1fd41c-b9f1-4008-8e17-3043627bbe72-kube-api-access-pgjpx\") pod \"redhat-marketplace-j9g9c\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:11 crc kubenswrapper[4912]: I1203 02:36:11.173415 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:11 crc kubenswrapper[4912]: I1203 02:36:11.880034 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9g9c"] Dec 03 02:36:12 crc kubenswrapper[4912]: I1203 02:36:12.829031 4912 generic.go:334] "Generic (PLEG): container finished" podID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerID="946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524" exitCode=0 Dec 03 02:36:12 crc kubenswrapper[4912]: I1203 02:36:12.829467 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9g9c" event={"ID":"2f1fd41c-b9f1-4008-8e17-3043627bbe72","Type":"ContainerDied","Data":"946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524"} Dec 03 02:36:12 crc kubenswrapper[4912]: I1203 02:36:12.829491 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9g9c" event={"ID":"2f1fd41c-b9f1-4008-8e17-3043627bbe72","Type":"ContainerStarted","Data":"31a69579a86bae9c3da0e204d4f5cf73829c7f22ed207087c7ff98fd375cac89"} Dec 03 02:36:13 crc kubenswrapper[4912]: I1203 02:36:13.802782 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e909a03f-b885-46d4-a139-833410cf9ff1/memcached/0.log" Dec 03 02:36:13 crc kubenswrapper[4912]: I1203 02:36:13.841451 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9g9c" event={"ID":"2f1fd41c-b9f1-4008-8e17-3043627bbe72","Type":"ContainerStarted","Data":"0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3"} Dec 03 02:36:14 crc kubenswrapper[4912]: I1203 02:36:14.859734 4912 generic.go:334] "Generic (PLEG): container finished" podID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerID="0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3" exitCode=0 Dec 03 02:36:14 crc kubenswrapper[4912]: I1203 02:36:14.859778 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9g9c" event={"ID":"2f1fd41c-b9f1-4008-8e17-3043627bbe72","Type":"ContainerDied","Data":"0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3"} Dec 03 02:36:15 crc kubenswrapper[4912]: I1203 02:36:15.871337 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9g9c" event={"ID":"2f1fd41c-b9f1-4008-8e17-3043627bbe72","Type":"ContainerStarted","Data":"fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3"} Dec 03 02:36:15 crc kubenswrapper[4912]: I1203 02:36:15.901074 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j9g9c" podStartSLOduration=3.420898917 podStartE2EDuration="5.901056637s" podCreationTimestamp="2025-12-03 02:36:10 +0000 UTC" firstStartedPulling="2025-12-03 02:36:12.831133902 +0000 UTC m=+7958.473154462" lastFinishedPulling="2025-12-03 02:36:15.311291622 +0000 UTC m=+7960.953312182" observedRunningTime="2025-12-03 02:36:15.893424903 +0000 UTC m=+7961.535445463" watchObservedRunningTime="2025-12-03 02:36:15.901056637 +0000 UTC m=+7961.543077197" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.430345 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p68mg"] Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.434258 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.448931 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p68mg"] Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.599712 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-catalog-content\") pod \"community-operators-p68mg\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.599929 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-utilities\") pod \"community-operators-p68mg\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.599976 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqmp7\" (UniqueName: \"kubernetes.io/projected/55e142a9-5cb4-462d-a803-4dfe0b9e3955-kube-api-access-nqmp7\") pod \"community-operators-p68mg\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.715994 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-utilities\") pod \"community-operators-p68mg\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.716100 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqmp7\" (UniqueName: \"kubernetes.io/projected/55e142a9-5cb4-462d-a803-4dfe0b9e3955-kube-api-access-nqmp7\") pod \"community-operators-p68mg\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.716403 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-catalog-content\") pod \"community-operators-p68mg\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.717204 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-catalog-content\") pod \"community-operators-p68mg\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.718504 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-utilities\") pod \"community-operators-p68mg\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.747197 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqmp7\" (UniqueName: \"kubernetes.io/projected/55e142a9-5cb4-462d-a803-4dfe0b9e3955-kube-api-access-nqmp7\") pod \"community-operators-p68mg\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:19 crc kubenswrapper[4912]: I1203 02:36:19.760657 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:20 crc kubenswrapper[4912]: W1203 02:36:20.312797 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55e142a9_5cb4_462d_a803_4dfe0b9e3955.slice/crio-d98189dc422b1e3433463ea6f7eaafedf31b674804c8c91fd962e99ec407b4b3 WatchSource:0}: Error finding container d98189dc422b1e3433463ea6f7eaafedf31b674804c8c91fd962e99ec407b4b3: Status 404 returned error can't find the container with id d98189dc422b1e3433463ea6f7eaafedf31b674804c8c91fd962e99ec407b4b3 Dec 03 02:36:20 crc kubenswrapper[4912]: I1203 02:36:20.313209 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p68mg"] Dec 03 02:36:20 crc kubenswrapper[4912]: I1203 02:36:20.573110 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:36:20 crc kubenswrapper[4912]: E1203 02:36:20.573623 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:36:20 crc kubenswrapper[4912]: I1203 02:36:20.942487 4912 generic.go:334] "Generic (PLEG): container finished" podID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerID="774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb" exitCode=0 Dec 03 02:36:20 crc kubenswrapper[4912]: I1203 02:36:20.942539 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p68mg" event={"ID":"55e142a9-5cb4-462d-a803-4dfe0b9e3955","Type":"ContainerDied","Data":"774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb"} Dec 03 02:36:20 crc kubenswrapper[4912]: I1203 02:36:20.942878 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p68mg" event={"ID":"55e142a9-5cb4-462d-a803-4dfe0b9e3955","Type":"ContainerStarted","Data":"d98189dc422b1e3433463ea6f7eaafedf31b674804c8c91fd962e99ec407b4b3"} Dec 03 02:36:21 crc kubenswrapper[4912]: I1203 02:36:21.174497 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:21 crc kubenswrapper[4912]: I1203 02:36:21.174554 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:21 crc kubenswrapper[4912]: I1203 02:36:21.241616 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:22 crc kubenswrapper[4912]: I1203 02:36:22.014246 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:22 crc kubenswrapper[4912]: I1203 02:36:22.971644 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p68mg" event={"ID":"55e142a9-5cb4-462d-a803-4dfe0b9e3955","Type":"ContainerStarted","Data":"4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479"} Dec 03 02:36:23 crc kubenswrapper[4912]: I1203 02:36:23.986214 4912 generic.go:334] "Generic (PLEG): container finished" podID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerID="4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479" exitCode=0 Dec 03 02:36:23 crc kubenswrapper[4912]: I1203 02:36:23.986319 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p68mg" event={"ID":"55e142a9-5cb4-462d-a803-4dfe0b9e3955","Type":"ContainerDied","Data":"4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479"} Dec 03 02:36:24 crc kubenswrapper[4912]: I1203 02:36:24.006059 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9g9c"] Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.001021 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p68mg" event={"ID":"55e142a9-5cb4-462d-a803-4dfe0b9e3955","Type":"ContainerStarted","Data":"0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207"} Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.001247 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j9g9c" podUID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerName="registry-server" containerID="cri-o://fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3" gracePeriod=2 Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.034896 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p68mg" podStartSLOduration=2.377335391 podStartE2EDuration="6.034874673s" podCreationTimestamp="2025-12-03 02:36:19 +0000 UTC" firstStartedPulling="2025-12-03 02:36:20.945207569 +0000 UTC m=+7966.587228129" lastFinishedPulling="2025-12-03 02:36:24.602746811 +0000 UTC m=+7970.244767411" observedRunningTime="2025-12-03 02:36:25.027564756 +0000 UTC m=+7970.669585326" watchObservedRunningTime="2025-12-03 02:36:25.034874673 +0000 UTC m=+7970.676895243" Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.567675 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.654080 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgjpx\" (UniqueName: \"kubernetes.io/projected/2f1fd41c-b9f1-4008-8e17-3043627bbe72-kube-api-access-pgjpx\") pod \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.654133 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-utilities\") pod \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.654536 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-catalog-content\") pod \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\" (UID: \"2f1fd41c-b9f1-4008-8e17-3043627bbe72\") " Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.656792 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-utilities" (OuterVolumeSpecName: "utilities") pod "2f1fd41c-b9f1-4008-8e17-3043627bbe72" (UID: "2f1fd41c-b9f1-4008-8e17-3043627bbe72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.664323 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f1fd41c-b9f1-4008-8e17-3043627bbe72-kube-api-access-pgjpx" (OuterVolumeSpecName: "kube-api-access-pgjpx") pod "2f1fd41c-b9f1-4008-8e17-3043627bbe72" (UID: "2f1fd41c-b9f1-4008-8e17-3043627bbe72"). InnerVolumeSpecName "kube-api-access-pgjpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.684748 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f1fd41c-b9f1-4008-8e17-3043627bbe72" (UID: "2f1fd41c-b9f1-4008-8e17-3043627bbe72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.757235 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.757278 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgjpx\" (UniqueName: \"kubernetes.io/projected/2f1fd41c-b9f1-4008-8e17-3043627bbe72-kube-api-access-pgjpx\") on node \"crc\" DevicePath \"\"" Dec 03 02:36:25 crc kubenswrapper[4912]: I1203 02:36:25.757291 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f1fd41c-b9f1-4008-8e17-3043627bbe72-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.013652 4912 generic.go:334] "Generic (PLEG): container finished" podID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerID="fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3" exitCode=0 Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.013738 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j9g9c" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.015545 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9g9c" event={"ID":"2f1fd41c-b9f1-4008-8e17-3043627bbe72","Type":"ContainerDied","Data":"fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3"} Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.015804 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j9g9c" event={"ID":"2f1fd41c-b9f1-4008-8e17-3043627bbe72","Type":"ContainerDied","Data":"31a69579a86bae9c3da0e204d4f5cf73829c7f22ed207087c7ff98fd375cac89"} Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.015894 4912 scope.go:117] "RemoveContainer" containerID="fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.063972 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9g9c"] Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.070453 4912 scope.go:117] "RemoveContainer" containerID="0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.097543 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j9g9c"] Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.098590 4912 scope.go:117] "RemoveContainer" containerID="946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.149767 4912 scope.go:117] "RemoveContainer" containerID="fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3" Dec 03 02:36:26 crc kubenswrapper[4912]: E1203 02:36:26.150385 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3\": container with ID starting with fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3 not found: ID does not exist" containerID="fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.150419 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3"} err="failed to get container status \"fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3\": rpc error: code = NotFound desc = could not find container \"fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3\": container with ID starting with fa62308fe4a9eadb0d855a9aac68afa33edff0bf04ca355217e54513232d3ae3 not found: ID does not exist" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.150457 4912 scope.go:117] "RemoveContainer" containerID="0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3" Dec 03 02:36:26 crc kubenswrapper[4912]: E1203 02:36:26.150648 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3\": container with ID starting with 0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3 not found: ID does not exist" containerID="0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.150669 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3"} err="failed to get container status \"0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3\": rpc error: code = NotFound desc = could not find container \"0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3\": container with ID starting with 0d0b5889b3d453fe18e85e8c5476dfa09633245e20387b2684eca9cfe21262b3 not found: ID does not exist" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.150685 4912 scope.go:117] "RemoveContainer" containerID="946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524" Dec 03 02:36:26 crc kubenswrapper[4912]: E1203 02:36:26.150849 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524\": container with ID starting with 946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524 not found: ID does not exist" containerID="946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.150871 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524"} err="failed to get container status \"946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524\": rpc error: code = NotFound desc = could not find container \"946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524\": container with ID starting with 946cc90cdc73c1dbd0dee3a10a56b28558dbdababc96bc8d2b3e407da4d07524 not found: ID does not exist" Dec 03 02:36:26 crc kubenswrapper[4912]: I1203 02:36:26.588940 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" path="/var/lib/kubelet/pods/2f1fd41c-b9f1-4008-8e17-3043627bbe72/volumes" Dec 03 02:36:29 crc kubenswrapper[4912]: I1203 02:36:29.761875 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:29 crc kubenswrapper[4912]: I1203 02:36:29.762779 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:29 crc kubenswrapper[4912]: I1203 02:36:29.858611 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:30 crc kubenswrapper[4912]: I1203 02:36:30.118846 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:31 crc kubenswrapper[4912]: I1203 02:36:31.003564 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p68mg"] Dec 03 02:36:31 crc kubenswrapper[4912]: I1203 02:36:31.817563 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-twd26_56cdea32-a9a4-410b-9a9c-c237746b0582/kube-rbac-proxy/0.log" Dec 03 02:36:31 crc kubenswrapper[4912]: I1203 02:36:31.923708 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-twd26_56cdea32-a9a4-410b-9a9c-c237746b0582/manager/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.048828 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-ds5mt_70a932ec-65ac-4616-870b-9297bd5d0a28/kube-rbac-proxy/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.083603 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p68mg" podUID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerName="registry-server" containerID="cri-o://0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207" gracePeriod=2 Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.143244 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-ds5mt_70a932ec-65ac-4616-870b-9297bd5d0a28/manager/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.292035 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-tqphq_5118703e-2fdb-41d6-a76e-ab5d3f6adb95/kube-rbac-proxy/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.366179 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-tqphq_5118703e-2fdb-41d6-a76e-ab5d3f6adb95/manager/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.426580 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/util/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.597843 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:36:32 crc kubenswrapper[4912]: E1203 02:36:32.598136 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.600724 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/util/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.617402 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.658140 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/pull/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.686609 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/pull/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.743408 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-catalog-content\") pod \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.743670 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-utilities\") pod \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.743730 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqmp7\" (UniqueName: \"kubernetes.io/projected/55e142a9-5cb4-462d-a803-4dfe0b9e3955-kube-api-access-nqmp7\") pod \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\" (UID: \"55e142a9-5cb4-462d-a803-4dfe0b9e3955\") " Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.744982 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-utilities" (OuterVolumeSpecName: "utilities") pod "55e142a9-5cb4-462d-a803-4dfe0b9e3955" (UID: "55e142a9-5cb4-462d-a803-4dfe0b9e3955"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.746982 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.758314 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e142a9-5cb4-462d-a803-4dfe0b9e3955-kube-api-access-nqmp7" (OuterVolumeSpecName: "kube-api-access-nqmp7") pod "55e142a9-5cb4-462d-a803-4dfe0b9e3955" (UID: "55e142a9-5cb4-462d-a803-4dfe0b9e3955"). InnerVolumeSpecName "kube-api-access-nqmp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.788993 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55e142a9-5cb4-462d-a803-4dfe0b9e3955" (UID: "55e142a9-5cb4-462d-a803-4dfe0b9e3955"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.849204 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqmp7\" (UniqueName: \"kubernetes.io/projected/55e142a9-5cb4-462d-a803-4dfe0b9e3955-kube-api-access-nqmp7\") on node \"crc\" DevicePath \"\"" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.849235 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55e142a9-5cb4-462d-a803-4dfe0b9e3955-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.851914 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/pull/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.853279 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/extract/0.log" Dec 03 02:36:32 crc kubenswrapper[4912]: I1203 02:36:32.860719 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/util/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.072180 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-g4n8g_d2643e7f-8db2-4374-b0d9-05a2130795aa/kube-rbac-proxy/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.095924 4912 generic.go:334] "Generic (PLEG): container finished" podID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerID="0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207" exitCode=0 Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.095957 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p68mg" event={"ID":"55e142a9-5cb4-462d-a803-4dfe0b9e3955","Type":"ContainerDied","Data":"0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207"} Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.095984 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p68mg" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.096150 4912 scope.go:117] "RemoveContainer" containerID="0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.096384 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p68mg" event={"ID":"55e142a9-5cb4-462d-a803-4dfe0b9e3955","Type":"ContainerDied","Data":"d98189dc422b1e3433463ea6f7eaafedf31b674804c8c91fd962e99ec407b4b3"} Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.109386 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-g4n8g_d2643e7f-8db2-4374-b0d9-05a2130795aa/manager/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.138270 4912 scope.go:117] "RemoveContainer" containerID="4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.151731 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p68mg"] Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.160518 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p68mg"] Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.170704 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jskq6_1df68f98-9a63-43ca-8e47-5c34c1e52581/kube-rbac-proxy/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.171196 4912 scope.go:117] "RemoveContainer" containerID="774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.226414 4912 scope.go:117] "RemoveContainer" containerID="0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207" Dec 03 02:36:33 crc kubenswrapper[4912]: E1203 02:36:33.226913 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207\": container with ID starting with 0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207 not found: ID does not exist" containerID="0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.226955 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207"} err="failed to get container status \"0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207\": rpc error: code = NotFound desc = could not find container \"0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207\": container with ID starting with 0253fa6da2ca0e49da69d4db882c40f9ba0a96c340b4b58349fc7dbeadfdc207 not found: ID does not exist" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.226980 4912 scope.go:117] "RemoveContainer" containerID="4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479" Dec 03 02:36:33 crc kubenswrapper[4912]: E1203 02:36:33.227296 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479\": container with ID starting with 4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479 not found: ID does not exist" containerID="4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.227335 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479"} err="failed to get container status \"4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479\": rpc error: code = NotFound desc = could not find container \"4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479\": container with ID starting with 4b948c7162dfd4bbaff984165b1de506a267b50ca893f74a1d433b05d3fda479 not found: ID does not exist" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.227348 4912 scope.go:117] "RemoveContainer" containerID="774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb" Dec 03 02:36:33 crc kubenswrapper[4912]: E1203 02:36:33.227558 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb\": container with ID starting with 774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb not found: ID does not exist" containerID="774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.227573 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb"} err="failed to get container status \"774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb\": rpc error: code = NotFound desc = could not find container \"774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb\": container with ID starting with 774eb18085ba38e6d74b20c080a65bed3570ee78b0b3905eeb44571e2917bcbb not found: ID does not exist" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.390226 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jskq6_1df68f98-9a63-43ca-8e47-5c34c1e52581/manager/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.412773 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-rb4z7_27605b80-0a7f-4b6b-9729-d0b6eaa0a74b/manager/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.433166 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-rb4z7_27605b80-0a7f-4b6b-9729-d0b6eaa0a74b/kube-rbac-proxy/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.541160 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-lzff2_045ab366-9392-40f9-94f5-95c983f65176/kube-rbac-proxy/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.733918 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-lzff2_045ab366-9392-40f9-94f5-95c983f65176/manager/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.768520 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-626ss_7b515331-ceab-4f45-9880-719c72dfcc4c/manager/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.804555 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-626ss_7b515331-ceab-4f45-9880-719c72dfcc4c/kube-rbac-proxy/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.897685 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-txtqb_9b23d447-fdc8-47a1-af38-a569b6ad2a8f/kube-rbac-proxy/0.log" Dec 03 02:36:33 crc kubenswrapper[4912]: I1203 02:36:33.996308 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-txtqb_9b23d447-fdc8-47a1-af38-a569b6ad2a8f/manager/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.049537 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-fxhh2_8ddf2fc1-446c-4a26-987a-f67ec4c51455/kube-rbac-proxy/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.138149 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-fxhh2_8ddf2fc1-446c-4a26-987a-f67ec4c51455/manager/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.389734 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-rjqjr_a7db7741-5afc-46bc-84c4-71c233ca9dde/manager/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.406697 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-rjqjr_a7db7741-5afc-46bc-84c4-71c233ca9dde/kube-rbac-proxy/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.483015 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-mjphw_6848c014-0383-482f-9d46-232cc8fafbe1/kube-rbac-proxy/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.606276 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-mjphw_6848c014-0383-482f-9d46-232cc8fafbe1/manager/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.606394 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" path="/var/lib/kubelet/pods/55e142a9-5cb4-462d-a803-4dfe0b9e3955/volumes" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.665669 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-bdgdj_de517f42-de01-4dde-ac22-1a651a38926b/kube-rbac-proxy/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.799558 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-bdgdj_de517f42-de01-4dde-ac22-1a651a38926b/manager/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.833303 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dtgrm_baaa7298-1635-4de5-af87-68db95102082/kube-rbac-proxy/0.log" Dec 03 02:36:34 crc kubenswrapper[4912]: I1203 02:36:34.904006 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dtgrm_baaa7298-1635-4de5-af87-68db95102082/manager/0.log" Dec 03 02:36:35 crc kubenswrapper[4912]: I1203 02:36:35.049280 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb_3e85fc81-3fbd-4af2-8c98-89892a4f3579/kube-rbac-proxy/0.log" Dec 03 02:36:35 crc kubenswrapper[4912]: I1203 02:36:35.077576 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb_3e85fc81-3fbd-4af2-8c98-89892a4f3579/manager/0.log" Dec 03 02:36:35 crc kubenswrapper[4912]: I1203 02:36:35.554404 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-c796f8559-s567r_e96cfd80-56de-4ab0-98b0-8bdcb324e498/operator/0.log" Dec 03 02:36:35 crc kubenswrapper[4912]: I1203 02:36:35.599466 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-4nzrg_5d286a0d-a818-4cb6-a798-135fda9d947e/registry-server/0.log" Dec 03 02:36:35 crc kubenswrapper[4912]: I1203 02:36:35.796524 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-vw7dk_fd3cb745-7c4b-484b-93a5-7638dcd4d738/kube-rbac-proxy/0.log" Dec 03 02:36:35 crc kubenswrapper[4912]: I1203 02:36:35.846902 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-vw7dk_fd3cb745-7c4b-484b-93a5-7638dcd4d738/manager/0.log" Dec 03 02:36:35 crc kubenswrapper[4912]: I1203 02:36:35.960510 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-5n67v_7ec29496-d0ac-42a2-acea-7037dfde41fb/kube-rbac-proxy/0.log" Dec 03 02:36:36 crc kubenswrapper[4912]: I1203 02:36:36.071848 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-5n67v_7ec29496-d0ac-42a2-acea-7037dfde41fb/manager/0.log" Dec 03 02:36:36 crc kubenswrapper[4912]: I1203 02:36:36.144460 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-llhvt_b54a05e7-9cfd-401d-8e5d-87c8ed5ad321/operator/0.log" Dec 03 02:36:36 crc kubenswrapper[4912]: I1203 02:36:36.333143 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-8x84c_2ab18e2a-3358-44e6-ac55-80b072619e73/kube-rbac-proxy/0.log" Dec 03 02:36:36 crc kubenswrapper[4912]: I1203 02:36:36.342670 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-8x84c_2ab18e2a-3358-44e6-ac55-80b072619e73/manager/0.log" Dec 03 02:36:36 crc kubenswrapper[4912]: I1203 02:36:36.579807 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d66d5c957-xgv9t_82b74254-de20-48a4-bfae-21afed7954cd/kube-rbac-proxy/0.log" Dec 03 02:36:36 crc kubenswrapper[4912]: I1203 02:36:36.831847 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-7rstc_b5544f39-6c3a-4737-a8e3-389a6e66ba40/manager/0.log" Dec 03 02:36:36 crc kubenswrapper[4912]: I1203 02:36:36.845571 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-7rstc_b5544f39-6c3a-4737-a8e3-389a6e66ba40/kube-rbac-proxy/0.log" Dec 03 02:36:36 crc kubenswrapper[4912]: I1203 02:36:36.974944 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86464467f8-sq576_10227b7f-ff98-4f4a-862f-3ef78b4e7737/manager/0.log" Dec 03 02:36:37 crc kubenswrapper[4912]: I1203 02:36:37.024544 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d66d5c957-xgv9t_82b74254-de20-48a4-bfae-21afed7954cd/manager/0.log" Dec 03 02:36:37 crc kubenswrapper[4912]: I1203 02:36:37.054007 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-jqzw7_ea0ec30d-2c13-49fd-883b-767024ff632c/kube-rbac-proxy/0.log" Dec 03 02:36:37 crc kubenswrapper[4912]: I1203 02:36:37.097839 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-jqzw7_ea0ec30d-2c13-49fd-883b-767024ff632c/manager/0.log" Dec 03 02:36:44 crc kubenswrapper[4912]: I1203 02:36:44.584800 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:36:44 crc kubenswrapper[4912]: E1203 02:36:44.585603 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:36:57 crc kubenswrapper[4912]: I1203 02:36:57.572248 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:36:57 crc kubenswrapper[4912]: E1203 02:36:57.572971 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:36:58 crc kubenswrapper[4912]: I1203 02:36:58.346285 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-g5h6h_cff6f92d-904b-43ba-a0d7-2bae917cdcd8/control-plane-machine-set-operator/0.log" Dec 03 02:36:58 crc kubenswrapper[4912]: I1203 02:36:58.509946 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tqgnx_ac6ae3d3-462a-4a82-a4a6-e225f12ce343/kube-rbac-proxy/0.log" Dec 03 02:36:58 crc kubenswrapper[4912]: I1203 02:36:58.574611 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tqgnx_ac6ae3d3-462a-4a82-a4a6-e225f12ce343/machine-api-operator/0.log" Dec 03 02:37:12 crc kubenswrapper[4912]: I1203 02:37:12.448095 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9hkxj_52072628-05d2-4d43-8fea-11ec571a1a5f/cert-manager-controller/0.log" Dec 03 02:37:12 crc kubenswrapper[4912]: I1203 02:37:12.575466 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:37:12 crc kubenswrapper[4912]: E1203 02:37:12.575758 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:37:12 crc kubenswrapper[4912]: I1203 02:37:12.602559 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-dlkbd_2620530c-1377-4da3-a983-a0211b396c18/cert-manager-cainjector/0.log" Dec 03 02:37:12 crc kubenswrapper[4912]: I1203 02:37:12.679004 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-s6b5l_77d9d66c-7ec2-4cc0-8240-bd02f4b710c7/cert-manager-webhook/0.log" Dec 03 02:37:25 crc kubenswrapper[4912]: I1203 02:37:25.571745 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:37:25 crc kubenswrapper[4912]: E1203 02:37:25.572588 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:37:26 crc kubenswrapper[4912]: I1203 02:37:26.522926 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-7dtpc_0c9a6fef-57f2-4add-99ff-2807455dd07d/nmstate-console-plugin/0.log" Dec 03 02:37:26 crc kubenswrapper[4912]: I1203 02:37:26.684251 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-667c7_af9b5997-afe3-45f8-8f6e-1b4b427c41cd/nmstate-handler/0.log" Dec 03 02:37:26 crc kubenswrapper[4912]: I1203 02:37:26.744757 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-w95jw_ff447976-7803-4a25-8491-f2094e165f86/nmstate-metrics/0.log" Dec 03 02:37:26 crc kubenswrapper[4912]: I1203 02:37:26.745287 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-w95jw_ff447976-7803-4a25-8491-f2094e165f86/kube-rbac-proxy/0.log" Dec 03 02:37:26 crc kubenswrapper[4912]: I1203 02:37:26.869396 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-mxpqr_48b78ad2-74d3-43ff-a69d-e729d3082199/nmstate-operator/0.log" Dec 03 02:37:26 crc kubenswrapper[4912]: I1203 02:37:26.936162 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-wt4wl_df373e95-5be1-45c8-a5a0-58dfd53caf3c/nmstate-webhook/0.log" Dec 03 02:37:37 crc kubenswrapper[4912]: I1203 02:37:37.571454 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:37:37 crc kubenswrapper[4912]: E1203 02:37:37.572223 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:37:40 crc kubenswrapper[4912]: I1203 02:37:40.452763 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-669f4ccbc6-wkghn_32b6d349-0f43-45f8-93d6-c5f9f66273b2/manager/0.log" Dec 03 02:37:40 crc kubenswrapper[4912]: I1203 02:37:40.476773 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-669f4ccbc6-wkghn_32b6d349-0f43-45f8-93d6-c5f9f66273b2/kube-rbac-proxy/0.log" Dec 03 02:37:52 crc kubenswrapper[4912]: I1203 02:37:52.572236 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:37:53 crc kubenswrapper[4912]: I1203 02:37:53.027512 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"04b59eb9656a1794c6f3aabe121548f59e54e26a3fb9718e163d6f7f9ded797c"} Dec 03 02:37:55 crc kubenswrapper[4912]: I1203 02:37:55.885462 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-kxsph_fe13031d-b775-459e-ad64-7445ca8cbe63/cluster-logging-operator/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.028557 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-dccc6_5d5921d6-eb7c-4430-b71f-104dc4fdff22/collector/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.101190 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_2ead2e0e-573f-4ef4-8122-34c91de2fabe/loki-compactor/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.188848 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-cr4z6_3aa94604-5389-4ef5-8e97-63167d5fb009/loki-distributor/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.264421 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-777c467b8d-tfg7h_00d2fcad-0e8b-4daf-ae2d-084456cd7144/opa/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.265346 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-777c467b8d-tfg7h_00d2fcad-0e8b-4daf-ae2d-084456cd7144/gateway/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.420220 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-777c467b8d-zsrn8_dd4ddd07-754f-48f9-b082-97b948a8d041/gateway/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.456411 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-777c467b8d-zsrn8_dd4ddd07-754f-48f9-b082-97b948a8d041/opa/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.554412 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_dfc8d0a7-a087-4dfb-a460-e63aeb64828e/loki-index-gateway/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.746668 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_33e70879-0df5-4529-969f-e88309e22cb0/loki-ingester/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.764891 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-vxq5n_3828ecda-60a7-4376-a5ec-f2730b95945d/loki-querier/0.log" Dec 03 02:37:56 crc kubenswrapper[4912]: I1203 02:37:56.898681 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-ljkc8_faaba064-9cc2-44fe-8516-29b31630da75/loki-query-frontend/0.log" Dec 03 02:38:11 crc kubenswrapper[4912]: I1203 02:38:11.753026 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-sqxw4_b4d86772-d2f0-4a24-80e7-339798fc2d2c/kube-rbac-proxy/0.log" Dec 03 02:38:11 crc kubenswrapper[4912]: I1203 02:38:11.959408 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-sqxw4_b4d86772-d2f0-4a24-80e7-339798fc2d2c/controller/0.log" Dec 03 02:38:11 crc kubenswrapper[4912]: I1203 02:38:11.998082 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-frr-files/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.173010 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-frr-files/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.211350 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-reloader/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.230623 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-metrics/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.270142 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-reloader/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.386390 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-frr-files/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.413693 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-reloader/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.447855 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-metrics/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.449591 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-metrics/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.615872 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/controller/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.619612 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-frr-files/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.633778 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-reloader/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.639971 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-metrics/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.815752 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/kube-rbac-proxy/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.854033 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/kube-rbac-proxy-frr/0.log" Dec 03 02:38:12 crc kubenswrapper[4912]: I1203 02:38:12.897572 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/frr-metrics/0.log" Dec 03 02:38:13 crc kubenswrapper[4912]: I1203 02:38:13.010758 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/reloader/0.log" Dec 03 02:38:13 crc kubenswrapper[4912]: I1203 02:38:13.165308 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-jlww2_9f8130d5-88a6-4a62-a381-e98a3207d9b3/frr-k8s-webhook-server/0.log" Dec 03 02:38:13 crc kubenswrapper[4912]: I1203 02:38:13.372211 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-76f7d69d98-sl5x6_ae8cc701-2530-4ef7-bd0c-596cb88ffc3a/manager/0.log" Dec 03 02:38:13 crc kubenswrapper[4912]: I1203 02:38:13.431735 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-58487cf9c5-5g5xb_cb334b7c-36eb-4863-a2ad-2f4f126a9354/webhook-server/0.log" Dec 03 02:38:13 crc kubenswrapper[4912]: I1203 02:38:13.654574 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9nt4p_bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36/kube-rbac-proxy/0.log" Dec 03 02:38:14 crc kubenswrapper[4912]: I1203 02:38:14.191215 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9nt4p_bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36/speaker/0.log" Dec 03 02:38:14 crc kubenswrapper[4912]: I1203 02:38:14.824297 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/frr/0.log" Dec 03 02:38:29 crc kubenswrapper[4912]: I1203 02:38:29.894672 4912 scope.go:117] "RemoveContainer" containerID="7b86baefc423c31ea916931508dba7de3292d139fb7a90441cd7da0c8fc1fb81" Dec 03 02:38:29 crc kubenswrapper[4912]: I1203 02:38:29.921260 4912 scope.go:117] "RemoveContainer" containerID="55e4eeafb54d38486052f752555e327363950ca380b81c7d5fa9202362ddcebd" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.002233 4912 scope.go:117] "RemoveContainer" containerID="6a3e1d78be372dc7650e89491772f673e3e667cee2c3ffa3427cd5e87e05ff53" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.061252 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/util/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.231180 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/pull/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.259008 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/util/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.280210 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/pull/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.375811 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/util/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.404398 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/pull/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.458849 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/extract/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.606716 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/util/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.788152 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/pull/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.790645 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/pull/0.log" Dec 03 02:38:30 crc kubenswrapper[4912]: I1203 02:38:30.814389 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/util/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.005016 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/pull/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.034421 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/util/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.073546 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/extract/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.194792 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/util/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.405108 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/pull/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.414385 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/util/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.494848 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/pull/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.687056 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/util/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.708698 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/pull/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.711524 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/extract/0.log" Dec 03 02:38:31 crc kubenswrapper[4912]: I1203 02:38:31.866618 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/util/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.085149 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/pull/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.091654 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/util/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.100692 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/pull/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.297110 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/util/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.371330 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/pull/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.417341 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/extract/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.539541 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/util/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.706937 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/util/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.721880 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/pull/0.log" Dec 03 02:38:32 crc kubenswrapper[4912]: I1203 02:38:32.754274 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/pull/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.026371 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/pull/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.027632 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/extract/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.084486 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/util/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.263362 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-utilities/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.406715 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-utilities/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.411546 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-content/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.412815 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-content/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.732534 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-utilities/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.774533 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-content/0.log" Dec 03 02:38:33 crc kubenswrapper[4912]: I1203 02:38:33.884776 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-utilities/0.log" Dec 03 02:38:34 crc kubenswrapper[4912]: I1203 02:38:34.077278 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-utilities/0.log" Dec 03 02:38:34 crc kubenswrapper[4912]: I1203 02:38:34.087238 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-content/0.log" Dec 03 02:38:34 crc kubenswrapper[4912]: I1203 02:38:34.154916 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-content/0.log" Dec 03 02:38:34 crc kubenswrapper[4912]: I1203 02:38:34.362029 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-content/0.log" Dec 03 02:38:34 crc kubenswrapper[4912]: I1203 02:38:34.508082 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-utilities/0.log" Dec 03 02:38:34 crc kubenswrapper[4912]: I1203 02:38:34.617534 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-24t69_f8dce103-6de5-4159-a1e3-4ae68c513ee7/marketplace-operator/0.log" Dec 03 02:38:34 crc kubenswrapper[4912]: I1203 02:38:34.820832 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-utilities/0.log" Dec 03 02:38:34 crc kubenswrapper[4912]: I1203 02:38:34.919275 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/registry-server/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.197874 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-content/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.219639 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-content/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.262458 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-utilities/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.317193 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-utilities/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.466340 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-content/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.579246 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-utilities/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.643819 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/registry-server/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.788822 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-content/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.821834 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-content/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.854976 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/registry-server/0.log" Dec 03 02:38:35 crc kubenswrapper[4912]: I1203 02:38:35.862730 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-utilities/0.log" Dec 03 02:38:36 crc kubenswrapper[4912]: I1203 02:38:36.109113 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-utilities/0.log" Dec 03 02:38:36 crc kubenswrapper[4912]: I1203 02:38:36.118283 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-content/0.log" Dec 03 02:38:36 crc kubenswrapper[4912]: I1203 02:38:36.754327 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/registry-server/0.log" Dec 03 02:38:51 crc kubenswrapper[4912]: I1203 02:38:51.244253 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-spkm8_377b37d8-070e-4463-9ea8-c698c4e289a4/prometheus-operator/0.log" Dec 03 02:38:51 crc kubenswrapper[4912]: I1203 02:38:51.419285 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0/prometheus-operator-admission-webhook/0.log" Dec 03 02:38:51 crc kubenswrapper[4912]: I1203 02:38:51.472647 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_194f5401-404a-43b8-b826-3df930390630/prometheus-operator-admission-webhook/0.log" Dec 03 02:38:51 crc kubenswrapper[4912]: I1203 02:38:51.620454 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-bsscx_f4a067c7-ef55-4185-8315-a61b5bfc8e22/operator/0.log" Dec 03 02:38:51 crc kubenswrapper[4912]: I1203 02:38:51.677260 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-96rq5_9d7ffd59-b442-43e2-b7b4-5a2fc42507d3/observability-ui-dashboards/0.log" Dec 03 02:38:51 crc kubenswrapper[4912]: I1203 02:38:51.819833 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-trfdr_1bba9178-9a01-47ac-9657-c277217019ca/perses-operator/0.log" Dec 03 02:39:07 crc kubenswrapper[4912]: I1203 02:39:07.107111 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-669f4ccbc6-wkghn_32b6d349-0f43-45f8-93d6-c5f9f66273b2/manager/0.log" Dec 03 02:39:07 crc kubenswrapper[4912]: I1203 02:39:07.111015 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-669f4ccbc6-wkghn_32b6d349-0f43-45f8-93d6-c5f9f66273b2/kube-rbac-proxy/0.log" Dec 03 02:39:18 crc kubenswrapper[4912]: E1203 02:39:18.490099 4912 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.20:59694->38.102.83.20:41775: write tcp 38.102.83.20:59694->38.102.83.20:41775: write: broken pipe Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.286255 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m425x"] Dec 03 02:40:00 crc kubenswrapper[4912]: E1203 02:40:00.287279 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerName="extract-content" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.287296 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerName="extract-content" Dec 03 02:40:00 crc kubenswrapper[4912]: E1203 02:40:00.287327 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerName="extract-utilities" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.287335 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerName="extract-utilities" Dec 03 02:40:00 crc kubenswrapper[4912]: E1203 02:40:00.287353 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerName="extract-content" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.287364 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerName="extract-content" Dec 03 02:40:00 crc kubenswrapper[4912]: E1203 02:40:00.287379 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerName="extract-utilities" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.287387 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerName="extract-utilities" Dec 03 02:40:00 crc kubenswrapper[4912]: E1203 02:40:00.287410 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerName="registry-server" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.287418 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerName="registry-server" Dec 03 02:40:00 crc kubenswrapper[4912]: E1203 02:40:00.291477 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerName="registry-server" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.291500 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerName="registry-server" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.291820 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="55e142a9-5cb4-462d-a803-4dfe0b9e3955" containerName="registry-server" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.291846 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f1fd41c-b9f1-4008-8e17-3043627bbe72" containerName="registry-server" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.293853 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.320022 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m425x"] Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.456325 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-utilities\") pod \"certified-operators-m425x\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.456577 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q95tg\" (UniqueName: \"kubernetes.io/projected/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-kube-api-access-q95tg\") pod \"certified-operators-m425x\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.456777 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-catalog-content\") pod \"certified-operators-m425x\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.558355 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q95tg\" (UniqueName: \"kubernetes.io/projected/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-kube-api-access-q95tg\") pod \"certified-operators-m425x\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.558504 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-catalog-content\") pod \"certified-operators-m425x\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.558601 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-utilities\") pod \"certified-operators-m425x\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.559244 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-utilities\") pod \"certified-operators-m425x\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.559286 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-catalog-content\") pod \"certified-operators-m425x\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.603772 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q95tg\" (UniqueName: \"kubernetes.io/projected/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-kube-api-access-q95tg\") pod \"certified-operators-m425x\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:00 crc kubenswrapper[4912]: I1203 02:40:00.636925 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:01 crc kubenswrapper[4912]: I1203 02:40:01.207126 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m425x"] Dec 03 02:40:01 crc kubenswrapper[4912]: I1203 02:40:01.627589 4912 generic.go:334] "Generic (PLEG): container finished" podID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerID="984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2" exitCode=0 Dec 03 02:40:01 crc kubenswrapper[4912]: I1203 02:40:01.627639 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m425x" event={"ID":"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c","Type":"ContainerDied","Data":"984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2"} Dec 03 02:40:01 crc kubenswrapper[4912]: I1203 02:40:01.627667 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m425x" event={"ID":"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c","Type":"ContainerStarted","Data":"215d7e9837e132afafb4efda07593a72fdc6e2308b4aa6738c3bb256cffca620"} Dec 03 02:40:01 crc kubenswrapper[4912]: I1203 02:40:01.630193 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 02:40:03 crc kubenswrapper[4912]: I1203 02:40:03.650589 4912 generic.go:334] "Generic (PLEG): container finished" podID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerID="0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0" exitCode=0 Dec 03 02:40:03 crc kubenswrapper[4912]: I1203 02:40:03.650639 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m425x" event={"ID":"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c","Type":"ContainerDied","Data":"0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0"} Dec 03 02:40:04 crc kubenswrapper[4912]: I1203 02:40:04.698924 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m425x" event={"ID":"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c","Type":"ContainerStarted","Data":"c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd"} Dec 03 02:40:04 crc kubenswrapper[4912]: I1203 02:40:04.740423 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m425x" podStartSLOduration=2.272122094 podStartE2EDuration="4.740403846s" podCreationTimestamp="2025-12-03 02:40:00 +0000 UTC" firstStartedPulling="2025-12-03 02:40:01.629961062 +0000 UTC m=+8187.271981622" lastFinishedPulling="2025-12-03 02:40:04.098242814 +0000 UTC m=+8189.740263374" observedRunningTime="2025-12-03 02:40:04.728739703 +0000 UTC m=+8190.370760263" watchObservedRunningTime="2025-12-03 02:40:04.740403846 +0000 UTC m=+8190.382424406" Dec 03 02:40:10 crc kubenswrapper[4912]: I1203 02:40:10.637574 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:10 crc kubenswrapper[4912]: I1203 02:40:10.638258 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:10 crc kubenswrapper[4912]: I1203 02:40:10.701564 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:10 crc kubenswrapper[4912]: I1203 02:40:10.841628 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:10 crc kubenswrapper[4912]: I1203 02:40:10.945940 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m425x"] Dec 03 02:40:12 crc kubenswrapper[4912]: I1203 02:40:12.793201 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m425x" podUID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerName="registry-server" containerID="cri-o://c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd" gracePeriod=2 Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.379315 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.489048 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q95tg\" (UniqueName: \"kubernetes.io/projected/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-kube-api-access-q95tg\") pod \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.489331 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-utilities\") pod \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.489381 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-catalog-content\") pod \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\" (UID: \"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c\") " Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.490980 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-utilities" (OuterVolumeSpecName: "utilities") pod "7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" (UID: "7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.515700 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-kube-api-access-q95tg" (OuterVolumeSpecName: "kube-api-access-q95tg") pod "7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" (UID: "7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c"). InnerVolumeSpecName "kube-api-access-q95tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.623497 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" (UID: "7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.625182 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q95tg\" (UniqueName: \"kubernetes.io/projected/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-kube-api-access-q95tg\") on node \"crc\" DevicePath \"\"" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.625275 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.625355 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.804095 4912 generic.go:334] "Generic (PLEG): container finished" podID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerID="c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd" exitCode=0 Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.804140 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m425x" event={"ID":"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c","Type":"ContainerDied","Data":"c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd"} Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.804166 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m425x" event={"ID":"7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c","Type":"ContainerDied","Data":"215d7e9837e132afafb4efda07593a72fdc6e2308b4aa6738c3bb256cffca620"} Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.804185 4912 scope.go:117] "RemoveContainer" containerID="c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.804223 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m425x" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.836023 4912 scope.go:117] "RemoveContainer" containerID="0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.845017 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m425x"] Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.853664 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m425x"] Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.858280 4912 scope.go:117] "RemoveContainer" containerID="984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.946972 4912 scope.go:117] "RemoveContainer" containerID="c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd" Dec 03 02:40:13 crc kubenswrapper[4912]: E1203 02:40:13.947522 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd\": container with ID starting with c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd not found: ID does not exist" containerID="c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.947567 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd"} err="failed to get container status \"c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd\": rpc error: code = NotFound desc = could not find container \"c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd\": container with ID starting with c2be43dcd1b5a7ea7df3c56f9bc8218d3580955f9eb73e14e0f86de9525921fd not found: ID does not exist" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.947594 4912 scope.go:117] "RemoveContainer" containerID="0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0" Dec 03 02:40:13 crc kubenswrapper[4912]: E1203 02:40:13.948320 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0\": container with ID starting with 0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0 not found: ID does not exist" containerID="0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.948391 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0"} err="failed to get container status \"0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0\": rpc error: code = NotFound desc = could not find container \"0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0\": container with ID starting with 0f900a189c30115446e7022765285abe278129081977cf5e088b97f7aec399f0 not found: ID does not exist" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.948421 4912 scope.go:117] "RemoveContainer" containerID="984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2" Dec 03 02:40:13 crc kubenswrapper[4912]: E1203 02:40:13.948864 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2\": container with ID starting with 984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2 not found: ID does not exist" containerID="984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2" Dec 03 02:40:13 crc kubenswrapper[4912]: I1203 02:40:13.948889 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2"} err="failed to get container status \"984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2\": rpc error: code = NotFound desc = could not find container \"984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2\": container with ID starting with 984363d615e275dce577345ec0ff59d50677aab0e58f0e6202fa3fe941a078d2 not found: ID does not exist" Dec 03 02:40:14 crc kubenswrapper[4912]: I1203 02:40:14.608169 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" path="/var/lib/kubelet/pods/7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c/volumes" Dec 03 02:40:18 crc kubenswrapper[4912]: I1203 02:40:18.078450 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:40:18 crc kubenswrapper[4912]: I1203 02:40:18.078710 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:40:30 crc kubenswrapper[4912]: I1203 02:40:30.127482 4912 scope.go:117] "RemoveContainer" containerID="24b107239d48a46b8f33ad1cbfe16feae80f1b2d7f54a7fb7c1b9d7701e94bbd" Dec 03 02:40:48 crc kubenswrapper[4912]: I1203 02:40:48.077848 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:40:48 crc kubenswrapper[4912]: I1203 02:40:48.078274 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:41:08 crc kubenswrapper[4912]: I1203 02:41:08.504347 4912 generic.go:334] "Generic (PLEG): container finished" podID="d96706ef-a068-4bab-bd26-97b7e80a8fde" containerID="7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4" exitCode=0 Dec 03 02:41:08 crc kubenswrapper[4912]: I1203 02:41:08.504450 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cj2nv/must-gather-49cx7" event={"ID":"d96706ef-a068-4bab-bd26-97b7e80a8fde","Type":"ContainerDied","Data":"7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4"} Dec 03 02:41:08 crc kubenswrapper[4912]: I1203 02:41:08.506191 4912 scope.go:117] "RemoveContainer" containerID="7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4" Dec 03 02:41:09 crc kubenswrapper[4912]: I1203 02:41:09.395330 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cj2nv_must-gather-49cx7_d96706ef-a068-4bab-bd26-97b7e80a8fde/gather/0.log" Dec 03 02:41:17 crc kubenswrapper[4912]: I1203 02:41:17.740743 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-cj2nv/must-gather-49cx7"] Dec 03 02:41:17 crc kubenswrapper[4912]: I1203 02:41:17.744191 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-cj2nv/must-gather-49cx7" podUID="d96706ef-a068-4bab-bd26-97b7e80a8fde" containerName="copy" containerID="cri-o://e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7" gracePeriod=2 Dec 03 02:41:17 crc kubenswrapper[4912]: I1203 02:41:17.758031 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-cj2nv/must-gather-49cx7"] Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.078298 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.078576 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.078622 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.079406 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04b59eb9656a1794c6f3aabe121548f59e54e26a3fb9718e163d6f7f9ded797c"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.079479 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://04b59eb9656a1794c6f3aabe121548f59e54e26a3fb9718e163d6f7f9ded797c" gracePeriod=600 Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.370315 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cj2nv_must-gather-49cx7_d96706ef-a068-4bab-bd26-97b7e80a8fde/copy/0.log" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.371017 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.512372 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lv8f\" (UniqueName: \"kubernetes.io/projected/d96706ef-a068-4bab-bd26-97b7e80a8fde-kube-api-access-8lv8f\") pod \"d96706ef-a068-4bab-bd26-97b7e80a8fde\" (UID: \"d96706ef-a068-4bab-bd26-97b7e80a8fde\") " Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.512527 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d96706ef-a068-4bab-bd26-97b7e80a8fde-must-gather-output\") pod \"d96706ef-a068-4bab-bd26-97b7e80a8fde\" (UID: \"d96706ef-a068-4bab-bd26-97b7e80a8fde\") " Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.518648 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d96706ef-a068-4bab-bd26-97b7e80a8fde-kube-api-access-8lv8f" (OuterVolumeSpecName: "kube-api-access-8lv8f") pod "d96706ef-a068-4bab-bd26-97b7e80a8fde" (UID: "d96706ef-a068-4bab-bd26-97b7e80a8fde"). InnerVolumeSpecName "kube-api-access-8lv8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.618228 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lv8f\" (UniqueName: \"kubernetes.io/projected/d96706ef-a068-4bab-bd26-97b7e80a8fde-kube-api-access-8lv8f\") on node \"crc\" DevicePath \"\"" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.643617 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-cj2nv_must-gather-49cx7_d96706ef-a068-4bab-bd26-97b7e80a8fde/copy/0.log" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.644561 4912 generic.go:334] "Generic (PLEG): container finished" podID="d96706ef-a068-4bab-bd26-97b7e80a8fde" containerID="e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7" exitCode=143 Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.644679 4912 scope.go:117] "RemoveContainer" containerID="e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.644845 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cj2nv/must-gather-49cx7" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.652803 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="04b59eb9656a1794c6f3aabe121548f59e54e26a3fb9718e163d6f7f9ded797c" exitCode=0 Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.652946 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"04b59eb9656a1794c6f3aabe121548f59e54e26a3fb9718e163d6f7f9ded797c"} Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.653009 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee"} Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.686308 4912 scope.go:117] "RemoveContainer" containerID="7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.744580 4912 scope.go:117] "RemoveContainer" containerID="e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7" Dec 03 02:41:18 crc kubenswrapper[4912]: E1203 02:41:18.746848 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7\": container with ID starting with e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7 not found: ID does not exist" containerID="e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.746879 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7"} err="failed to get container status \"e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7\": rpc error: code = NotFound desc = could not find container \"e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7\": container with ID starting with e4901f19188fa7a7d4f54ca39eb2c6980326c55602a8408cbda163ab2c1880d7 not found: ID does not exist" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.746901 4912 scope.go:117] "RemoveContainer" containerID="7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4" Dec 03 02:41:18 crc kubenswrapper[4912]: E1203 02:41:18.747184 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4\": container with ID starting with 7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4 not found: ID does not exist" containerID="7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.747203 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4"} err="failed to get container status \"7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4\": rpc error: code = NotFound desc = could not find container \"7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4\": container with ID starting with 7a02505628fe317ce97ec718e986137fbb90e3e83ba3c81775db3b0ecddd91e4 not found: ID does not exist" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.747217 4912 scope.go:117] "RemoveContainer" containerID="5b24dd8e4e0411e804c2073835156a20b3024d4cfcaf980f032b83b6939373bb" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.750001 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d96706ef-a068-4bab-bd26-97b7e80a8fde-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d96706ef-a068-4bab-bd26-97b7e80a8fde" (UID: "d96706ef-a068-4bab-bd26-97b7e80a8fde"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:41:18 crc kubenswrapper[4912]: I1203 02:41:18.829880 4912 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d96706ef-a068-4bab-bd26-97b7e80a8fde-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 02:41:20 crc kubenswrapper[4912]: I1203 02:41:20.598780 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d96706ef-a068-4bab-bd26-97b7e80a8fde" path="/var/lib/kubelet/pods/d96706ef-a068-4bab-bd26-97b7e80a8fde/volumes" Dec 03 02:41:30 crc kubenswrapper[4912]: I1203 02:41:30.228663 4912 scope.go:117] "RemoveContainer" containerID="70691aaef0b9b021629063f1fd4e166b881957e9a1f292acf3b029cef3c17031" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.961422 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qqcfw"] Dec 03 02:43:13 crc kubenswrapper[4912]: E1203 02:43:13.963379 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerName="registry-server" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.963481 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerName="registry-server" Dec 03 02:43:13 crc kubenswrapper[4912]: E1203 02:43:13.963591 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d96706ef-a068-4bab-bd26-97b7e80a8fde" containerName="gather" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.963610 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d96706ef-a068-4bab-bd26-97b7e80a8fde" containerName="gather" Dec 03 02:43:13 crc kubenswrapper[4912]: E1203 02:43:13.963650 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d96706ef-a068-4bab-bd26-97b7e80a8fde" containerName="copy" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.963666 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d96706ef-a068-4bab-bd26-97b7e80a8fde" containerName="copy" Dec 03 02:43:13 crc kubenswrapper[4912]: E1203 02:43:13.963719 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerName="extract-content" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.963735 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerName="extract-content" Dec 03 02:43:13 crc kubenswrapper[4912]: E1203 02:43:13.963768 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerName="extract-utilities" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.963784 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerName="extract-utilities" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.964309 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b9ca3e2-bdd7-4534-948c-91e6d7a55c4c" containerName="registry-server" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.964363 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d96706ef-a068-4bab-bd26-97b7e80a8fde" containerName="gather" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.964414 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d96706ef-a068-4bab-bd26-97b7e80a8fde" containerName="copy" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.967695 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:13 crc kubenswrapper[4912]: I1203 02:43:13.978453 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qqcfw"] Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.102541 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmm99\" (UniqueName: \"kubernetes.io/projected/15f4b8f1-417f-4587-b9e6-c3de4c091f94-kube-api-access-vmm99\") pod \"redhat-operators-qqcfw\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.102610 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-utilities\") pod \"redhat-operators-qqcfw\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.102910 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-catalog-content\") pod \"redhat-operators-qqcfw\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.204628 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-catalog-content\") pod \"redhat-operators-qqcfw\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.204709 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmm99\" (UniqueName: \"kubernetes.io/projected/15f4b8f1-417f-4587-b9e6-c3de4c091f94-kube-api-access-vmm99\") pod \"redhat-operators-qqcfw\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.204745 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-utilities\") pod \"redhat-operators-qqcfw\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.205258 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-catalog-content\") pod \"redhat-operators-qqcfw\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.205310 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-utilities\") pod \"redhat-operators-qqcfw\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.230005 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmm99\" (UniqueName: \"kubernetes.io/projected/15f4b8f1-417f-4587-b9e6-c3de4c091f94-kube-api-access-vmm99\") pod \"redhat-operators-qqcfw\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.297775 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:14 crc kubenswrapper[4912]: I1203 02:43:14.833018 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qqcfw"] Dec 03 02:43:15 crc kubenswrapper[4912]: I1203 02:43:15.190606 4912 generic.go:334] "Generic (PLEG): container finished" podID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerID="e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c" exitCode=0 Dec 03 02:43:15 crc kubenswrapper[4912]: I1203 02:43:15.190698 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqcfw" event={"ID":"15f4b8f1-417f-4587-b9e6-c3de4c091f94","Type":"ContainerDied","Data":"e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c"} Dec 03 02:43:15 crc kubenswrapper[4912]: I1203 02:43:15.190899 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqcfw" event={"ID":"15f4b8f1-417f-4587-b9e6-c3de4c091f94","Type":"ContainerStarted","Data":"d84b11a7db201f8de822dca94ed5b1b49975d82059c1e2ba177b07d16973fa8b"} Dec 03 02:43:17 crc kubenswrapper[4912]: I1203 02:43:17.212406 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqcfw" event={"ID":"15f4b8f1-417f-4587-b9e6-c3de4c091f94","Type":"ContainerStarted","Data":"e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262"} Dec 03 02:43:18 crc kubenswrapper[4912]: I1203 02:43:18.077983 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:43:18 crc kubenswrapper[4912]: I1203 02:43:18.078411 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:43:20 crc kubenswrapper[4912]: I1203 02:43:20.245855 4912 generic.go:334] "Generic (PLEG): container finished" podID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerID="e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262" exitCode=0 Dec 03 02:43:20 crc kubenswrapper[4912]: I1203 02:43:20.245921 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqcfw" event={"ID":"15f4b8f1-417f-4587-b9e6-c3de4c091f94","Type":"ContainerDied","Data":"e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262"} Dec 03 02:43:21 crc kubenswrapper[4912]: I1203 02:43:21.258539 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqcfw" event={"ID":"15f4b8f1-417f-4587-b9e6-c3de4c091f94","Type":"ContainerStarted","Data":"9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e"} Dec 03 02:43:21 crc kubenswrapper[4912]: I1203 02:43:21.284042 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qqcfw" podStartSLOduration=2.623586329 podStartE2EDuration="8.284022867s" podCreationTimestamp="2025-12-03 02:43:13 +0000 UTC" firstStartedPulling="2025-12-03 02:43:15.192555326 +0000 UTC m=+8380.834575886" lastFinishedPulling="2025-12-03 02:43:20.852991864 +0000 UTC m=+8386.495012424" observedRunningTime="2025-12-03 02:43:21.282179137 +0000 UTC m=+8386.924199717" watchObservedRunningTime="2025-12-03 02:43:21.284022867 +0000 UTC m=+8386.926043447" Dec 03 02:43:24 crc kubenswrapper[4912]: I1203 02:43:24.298108 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:24 crc kubenswrapper[4912]: I1203 02:43:24.298874 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:25 crc kubenswrapper[4912]: I1203 02:43:25.383190 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qqcfw" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerName="registry-server" probeResult="failure" output=< Dec 03 02:43:25 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 02:43:25 crc kubenswrapper[4912]: > Dec 03 02:43:34 crc kubenswrapper[4912]: I1203 02:43:34.403895 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:34 crc kubenswrapper[4912]: I1203 02:43:34.489696 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:34 crc kubenswrapper[4912]: I1203 02:43:34.666341 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qqcfw"] Dec 03 02:43:35 crc kubenswrapper[4912]: I1203 02:43:35.434874 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qqcfw" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerName="registry-server" containerID="cri-o://9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e" gracePeriod=2 Dec 03 02:43:35 crc kubenswrapper[4912]: E1203 02:43:35.634242 4912 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15f4b8f1_417f_4587_b9e6_c3de4c091f94.slice/crio-conmon-9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e.scope\": RecentStats: unable to find data in memory cache]" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.060427 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.126112 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-catalog-content\") pod \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.126355 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmm99\" (UniqueName: \"kubernetes.io/projected/15f4b8f1-417f-4587-b9e6-c3de4c091f94-kube-api-access-vmm99\") pod \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.126590 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-utilities\") pod \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\" (UID: \"15f4b8f1-417f-4587-b9e6-c3de4c091f94\") " Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.127043 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-utilities" (OuterVolumeSpecName: "utilities") pod "15f4b8f1-417f-4587-b9e6-c3de4c091f94" (UID: "15f4b8f1-417f-4587-b9e6-c3de4c091f94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.127696 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.133211 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15f4b8f1-417f-4587-b9e6-c3de4c091f94-kube-api-access-vmm99" (OuterVolumeSpecName: "kube-api-access-vmm99") pod "15f4b8f1-417f-4587-b9e6-c3de4c091f94" (UID: "15f4b8f1-417f-4587-b9e6-c3de4c091f94"). InnerVolumeSpecName "kube-api-access-vmm99". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.230380 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmm99\" (UniqueName: \"kubernetes.io/projected/15f4b8f1-417f-4587-b9e6-c3de4c091f94-kube-api-access-vmm99\") on node \"crc\" DevicePath \"\"" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.246946 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15f4b8f1-417f-4587-b9e6-c3de4c091f94" (UID: "15f4b8f1-417f-4587-b9e6-c3de4c091f94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.332290 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15f4b8f1-417f-4587-b9e6-c3de4c091f94-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.458295 4912 generic.go:334] "Generic (PLEG): container finished" podID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerID="9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e" exitCode=0 Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.458338 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqcfw" event={"ID":"15f4b8f1-417f-4587-b9e6-c3de4c091f94","Type":"ContainerDied","Data":"9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e"} Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.458364 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqcfw" event={"ID":"15f4b8f1-417f-4587-b9e6-c3de4c091f94","Type":"ContainerDied","Data":"d84b11a7db201f8de822dca94ed5b1b49975d82059c1e2ba177b07d16973fa8b"} Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.458383 4912 scope.go:117] "RemoveContainer" containerID="9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.458533 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqcfw" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.514406 4912 scope.go:117] "RemoveContainer" containerID="e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.545814 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qqcfw"] Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.548340 4912 scope.go:117] "RemoveContainer" containerID="e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.561521 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qqcfw"] Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.586320 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" path="/var/lib/kubelet/pods/15f4b8f1-417f-4587-b9e6-c3de4c091f94/volumes" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.622243 4912 scope.go:117] "RemoveContainer" containerID="9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e" Dec 03 02:43:36 crc kubenswrapper[4912]: E1203 02:43:36.622844 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e\": container with ID starting with 9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e not found: ID does not exist" containerID="9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.622902 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e"} err="failed to get container status \"9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e\": rpc error: code = NotFound desc = could not find container \"9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e\": container with ID starting with 9a5a2d1d9be38de028cba463cfb6ffb42c2d93ff98a8b3eb67af2b49165e334e not found: ID does not exist" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.622951 4912 scope.go:117] "RemoveContainer" containerID="e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262" Dec 03 02:43:36 crc kubenswrapper[4912]: E1203 02:43:36.626868 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262\": container with ID starting with e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262 not found: ID does not exist" containerID="e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.626914 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262"} err="failed to get container status \"e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262\": rpc error: code = NotFound desc = could not find container \"e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262\": container with ID starting with e1ba23c01d8c8909abfa6de8277024c4a0730f965cbf26809d3f0154e4d03262 not found: ID does not exist" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.626939 4912 scope.go:117] "RemoveContainer" containerID="e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c" Dec 03 02:43:36 crc kubenswrapper[4912]: E1203 02:43:36.628776 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c\": container with ID starting with e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c not found: ID does not exist" containerID="e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c" Dec 03 02:43:36 crc kubenswrapper[4912]: I1203 02:43:36.628815 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c"} err="failed to get container status \"e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c\": rpc error: code = NotFound desc = could not find container \"e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c\": container with ID starting with e782125fc10a76308fa460f0960c7ae4a3070652d5db97b4bb5eb552050d136c not found: ID does not exist" Dec 03 02:43:48 crc kubenswrapper[4912]: I1203 02:43:48.077765 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:43:48 crc kubenswrapper[4912]: I1203 02:43:48.080140 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:44:18 crc kubenswrapper[4912]: I1203 02:44:18.077982 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:44:18 crc kubenswrapper[4912]: I1203 02:44:18.078657 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:44:18 crc kubenswrapper[4912]: I1203 02:44:18.078725 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:44:18 crc kubenswrapper[4912]: I1203 02:44:18.080007 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:44:18 crc kubenswrapper[4912]: I1203 02:44:18.080100 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" gracePeriod=600 Dec 03 02:44:18 crc kubenswrapper[4912]: E1203 02:44:18.224527 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:44:19 crc kubenswrapper[4912]: I1203 02:44:19.012617 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" exitCode=0 Dec 03 02:44:19 crc kubenswrapper[4912]: I1203 02:44:19.012682 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee"} Dec 03 02:44:19 crc kubenswrapper[4912]: I1203 02:44:19.012730 4912 scope.go:117] "RemoveContainer" containerID="04b59eb9656a1794c6f3aabe121548f59e54e26a3fb9718e163d6f7f9ded797c" Dec 03 02:44:19 crc kubenswrapper[4912]: I1203 02:44:19.013565 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:44:19 crc kubenswrapper[4912]: E1203 02:44:19.014173 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:44:33 crc kubenswrapper[4912]: I1203 02:44:33.571381 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:44:33 crc kubenswrapper[4912]: E1203 02:44:33.572193 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:44:47 crc kubenswrapper[4912]: I1203 02:44:47.573044 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:44:47 crc kubenswrapper[4912]: E1203 02:44:47.574359 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:44:49 crc kubenswrapper[4912]: I1203 02:44:49.972340 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x4djm/must-gather-5qw4w"] Dec 03 02:44:49 crc kubenswrapper[4912]: E1203 02:44:49.973493 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerName="extract-content" Dec 03 02:44:49 crc kubenswrapper[4912]: I1203 02:44:49.973513 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerName="extract-content" Dec 03 02:44:49 crc kubenswrapper[4912]: E1203 02:44:49.973529 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerName="extract-utilities" Dec 03 02:44:49 crc kubenswrapper[4912]: I1203 02:44:49.973536 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerName="extract-utilities" Dec 03 02:44:49 crc kubenswrapper[4912]: E1203 02:44:49.973593 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerName="registry-server" Dec 03 02:44:49 crc kubenswrapper[4912]: I1203 02:44:49.973603 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerName="registry-server" Dec 03 02:44:49 crc kubenswrapper[4912]: I1203 02:44:49.973882 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="15f4b8f1-417f-4587-b9e6-c3de4c091f94" containerName="registry-server" Dec 03 02:44:49 crc kubenswrapper[4912]: I1203 02:44:49.975460 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:44:49 crc kubenswrapper[4912]: I1203 02:44:49.995598 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x4djm"/"kube-root-ca.crt" Dec 03 02:44:49 crc kubenswrapper[4912]: I1203 02:44:49.995819 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-x4djm"/"openshift-service-ca.crt" Dec 03 02:44:50 crc kubenswrapper[4912]: I1203 02:44:50.010505 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x4djm/must-gather-5qw4w"] Dec 03 02:44:50 crc kubenswrapper[4912]: I1203 02:44:50.147984 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2lhj\" (UniqueName: \"kubernetes.io/projected/ed036cee-58c4-4577-8785-c1903431dd3a-kube-api-access-f2lhj\") pod \"must-gather-5qw4w\" (UID: \"ed036cee-58c4-4577-8785-c1903431dd3a\") " pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:44:50 crc kubenswrapper[4912]: I1203 02:44:50.148240 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed036cee-58c4-4577-8785-c1903431dd3a-must-gather-output\") pod \"must-gather-5qw4w\" (UID: \"ed036cee-58c4-4577-8785-c1903431dd3a\") " pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:44:50 crc kubenswrapper[4912]: I1203 02:44:50.250679 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed036cee-58c4-4577-8785-c1903431dd3a-must-gather-output\") pod \"must-gather-5qw4w\" (UID: \"ed036cee-58c4-4577-8785-c1903431dd3a\") " pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:44:50 crc kubenswrapper[4912]: I1203 02:44:50.250818 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2lhj\" (UniqueName: \"kubernetes.io/projected/ed036cee-58c4-4577-8785-c1903431dd3a-kube-api-access-f2lhj\") pod \"must-gather-5qw4w\" (UID: \"ed036cee-58c4-4577-8785-c1903431dd3a\") " pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:44:50 crc kubenswrapper[4912]: I1203 02:44:50.251128 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed036cee-58c4-4577-8785-c1903431dd3a-must-gather-output\") pod \"must-gather-5qw4w\" (UID: \"ed036cee-58c4-4577-8785-c1903431dd3a\") " pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:44:50 crc kubenswrapper[4912]: I1203 02:44:50.270752 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2lhj\" (UniqueName: \"kubernetes.io/projected/ed036cee-58c4-4577-8785-c1903431dd3a-kube-api-access-f2lhj\") pod \"must-gather-5qw4w\" (UID: \"ed036cee-58c4-4577-8785-c1903431dd3a\") " pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:44:50 crc kubenswrapper[4912]: I1203 02:44:50.323167 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:44:50 crc kubenswrapper[4912]: I1203 02:44:50.797672 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-x4djm/must-gather-5qw4w"] Dec 03 02:44:50 crc kubenswrapper[4912]: W1203 02:44:50.800714 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded036cee_58c4_4577_8785_c1903431dd3a.slice/crio-52e9ce8e081f8b037523cf3f26d81395ae4ece039002e02b30390c3796e0edf9 WatchSource:0}: Error finding container 52e9ce8e081f8b037523cf3f26d81395ae4ece039002e02b30390c3796e0edf9: Status 404 returned error can't find the container with id 52e9ce8e081f8b037523cf3f26d81395ae4ece039002e02b30390c3796e0edf9 Dec 03 02:44:51 crc kubenswrapper[4912]: I1203 02:44:51.418773 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/must-gather-5qw4w" event={"ID":"ed036cee-58c4-4577-8785-c1903431dd3a","Type":"ContainerStarted","Data":"5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25"} Dec 03 02:44:51 crc kubenswrapper[4912]: I1203 02:44:51.419036 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/must-gather-5qw4w" event={"ID":"ed036cee-58c4-4577-8785-c1903431dd3a","Type":"ContainerStarted","Data":"52e9ce8e081f8b037523cf3f26d81395ae4ece039002e02b30390c3796e0edf9"} Dec 03 02:44:52 crc kubenswrapper[4912]: I1203 02:44:52.476738 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/must-gather-5qw4w" event={"ID":"ed036cee-58c4-4577-8785-c1903431dd3a","Type":"ContainerStarted","Data":"1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe"} Dec 03 02:44:52 crc kubenswrapper[4912]: I1203 02:44:52.498844 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x4djm/must-gather-5qw4w" podStartSLOduration=3.498825931 podStartE2EDuration="3.498825931s" podCreationTimestamp="2025-12-03 02:44:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 02:44:52.497267389 +0000 UTC m=+8478.139287959" watchObservedRunningTime="2025-12-03 02:44:52.498825931 +0000 UTC m=+8478.140846501" Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.386163 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x4djm/crc-debug-6rmxg"] Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.388970 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.393450 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x4djm"/"default-dockercfg-fhkb2" Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.520764 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmpd\" (UniqueName: \"kubernetes.io/projected/b7f11bf1-cada-4639-b216-dbd785813d62-kube-api-access-kgmpd\") pod \"crc-debug-6rmxg\" (UID: \"b7f11bf1-cada-4639-b216-dbd785813d62\") " pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.520876 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7f11bf1-cada-4639-b216-dbd785813d62-host\") pod \"crc-debug-6rmxg\" (UID: \"b7f11bf1-cada-4639-b216-dbd785813d62\") " pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.622784 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmpd\" (UniqueName: \"kubernetes.io/projected/b7f11bf1-cada-4639-b216-dbd785813d62-kube-api-access-kgmpd\") pod \"crc-debug-6rmxg\" (UID: \"b7f11bf1-cada-4639-b216-dbd785813d62\") " pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.623160 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7f11bf1-cada-4639-b216-dbd785813d62-host\") pod \"crc-debug-6rmxg\" (UID: \"b7f11bf1-cada-4639-b216-dbd785813d62\") " pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.623365 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7f11bf1-cada-4639-b216-dbd785813d62-host\") pod \"crc-debug-6rmxg\" (UID: \"b7f11bf1-cada-4639-b216-dbd785813d62\") " pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.647579 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmpd\" (UniqueName: \"kubernetes.io/projected/b7f11bf1-cada-4639-b216-dbd785813d62-kube-api-access-kgmpd\") pod \"crc-debug-6rmxg\" (UID: \"b7f11bf1-cada-4639-b216-dbd785813d62\") " pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:44:55 crc kubenswrapper[4912]: I1203 02:44:55.713286 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:44:55 crc kubenswrapper[4912]: W1203 02:44:55.752662 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7f11bf1_cada_4639_b216_dbd785813d62.slice/crio-9af55abd89254f7a441b324ac5ba532c891a10f112155138d6a50140ac12f54c WatchSource:0}: Error finding container 9af55abd89254f7a441b324ac5ba532c891a10f112155138d6a50140ac12f54c: Status 404 returned error can't find the container with id 9af55abd89254f7a441b324ac5ba532c891a10f112155138d6a50140ac12f54c Dec 03 02:44:56 crc kubenswrapper[4912]: I1203 02:44:56.523942 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/crc-debug-6rmxg" event={"ID":"b7f11bf1-cada-4639-b216-dbd785813d62","Type":"ContainerStarted","Data":"7f43699422378b073dcf66075391c2e53b1c36d621a513cb4192f1e6e964dd04"} Dec 03 02:44:56 crc kubenswrapper[4912]: I1203 02:44:56.524598 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/crc-debug-6rmxg" event={"ID":"b7f11bf1-cada-4639-b216-dbd785813d62","Type":"ContainerStarted","Data":"9af55abd89254f7a441b324ac5ba532c891a10f112155138d6a50140ac12f54c"} Dec 03 02:44:56 crc kubenswrapper[4912]: I1203 02:44:56.545858 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-x4djm/crc-debug-6rmxg" podStartSLOduration=1.5458372310000001 podStartE2EDuration="1.545837231s" podCreationTimestamp="2025-12-03 02:44:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 02:44:56.543009275 +0000 UTC m=+8482.185029835" watchObservedRunningTime="2025-12-03 02:44:56.545837231 +0000 UTC m=+8482.187857811" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.180757 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp"] Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.183082 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.185735 4912 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.193953 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.194733 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp"] Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.231565 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-secret-volume\") pod \"collect-profiles-29412165-567kp\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.231623 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-config-volume\") pod \"collect-profiles-29412165-567kp\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.231681 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4q7t\" (UniqueName: \"kubernetes.io/projected/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-kube-api-access-c4q7t\") pod \"collect-profiles-29412165-567kp\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.333136 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-secret-volume\") pod \"collect-profiles-29412165-567kp\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.333180 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-config-volume\") pod \"collect-profiles-29412165-567kp\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.333213 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4q7t\" (UniqueName: \"kubernetes.io/projected/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-kube-api-access-c4q7t\") pod \"collect-profiles-29412165-567kp\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.334470 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-config-volume\") pod \"collect-profiles-29412165-567kp\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.346495 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-secret-volume\") pod \"collect-profiles-29412165-567kp\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.349990 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4q7t\" (UniqueName: \"kubernetes.io/projected/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-kube-api-access-c4q7t\") pod \"collect-profiles-29412165-567kp\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:00 crc kubenswrapper[4912]: I1203 02:45:00.505108 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:01 crc kubenswrapper[4912]: I1203 02:45:01.037231 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp"] Dec 03 02:45:01 crc kubenswrapper[4912]: I1203 02:45:01.572525 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:45:01 crc kubenswrapper[4912]: E1203 02:45:01.573160 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:45:01 crc kubenswrapper[4912]: I1203 02:45:01.617339 4912 generic.go:334] "Generic (PLEG): container finished" podID="d14eb280-2992-4a7c-96b5-64e4e8d75c6f" containerID="60bf9ca14647fc64dcceddd89ff263333eb2411af965f73cbc70bd35b3b6ded1" exitCode=0 Dec 03 02:45:01 crc kubenswrapper[4912]: I1203 02:45:01.617377 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" event={"ID":"d14eb280-2992-4a7c-96b5-64e4e8d75c6f","Type":"ContainerDied","Data":"60bf9ca14647fc64dcceddd89ff263333eb2411af965f73cbc70bd35b3b6ded1"} Dec 03 02:45:01 crc kubenswrapper[4912]: I1203 02:45:01.617401 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" event={"ID":"d14eb280-2992-4a7c-96b5-64e4e8d75c6f","Type":"ContainerStarted","Data":"88b610c564a2d3ad6a0a41875f5324aad2709fe66880e4616febfb08d0f3ac09"} Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.039482 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.206751 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4q7t\" (UniqueName: \"kubernetes.io/projected/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-kube-api-access-c4q7t\") pod \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.207114 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-secret-volume\") pod \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.207147 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-config-volume\") pod \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\" (UID: \"d14eb280-2992-4a7c-96b5-64e4e8d75c6f\") " Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.208673 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-config-volume" (OuterVolumeSpecName: "config-volume") pod "d14eb280-2992-4a7c-96b5-64e4e8d75c6f" (UID: "d14eb280-2992-4a7c-96b5-64e4e8d75c6f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.220593 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d14eb280-2992-4a7c-96b5-64e4e8d75c6f" (UID: "d14eb280-2992-4a7c-96b5-64e4e8d75c6f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.220646 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-kube-api-access-c4q7t" (OuterVolumeSpecName: "kube-api-access-c4q7t") pod "d14eb280-2992-4a7c-96b5-64e4e8d75c6f" (UID: "d14eb280-2992-4a7c-96b5-64e4e8d75c6f"). InnerVolumeSpecName "kube-api-access-c4q7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.310498 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4q7t\" (UniqueName: \"kubernetes.io/projected/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-kube-api-access-c4q7t\") on node \"crc\" DevicePath \"\"" Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.310657 4912 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.310769 4912 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d14eb280-2992-4a7c-96b5-64e4e8d75c6f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.642416 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" event={"ID":"d14eb280-2992-4a7c-96b5-64e4e8d75c6f","Type":"ContainerDied","Data":"88b610c564a2d3ad6a0a41875f5324aad2709fe66880e4616febfb08d0f3ac09"} Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.642673 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88b610c564a2d3ad6a0a41875f5324aad2709fe66880e4616febfb08d0f3ac09" Dec 03 02:45:03 crc kubenswrapper[4912]: I1203 02:45:03.642727 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29412165-567kp" Dec 03 02:45:04 crc kubenswrapper[4912]: I1203 02:45:04.135182 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj"] Dec 03 02:45:04 crc kubenswrapper[4912]: I1203 02:45:04.146884 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29412120-p6phj"] Dec 03 02:45:04 crc kubenswrapper[4912]: I1203 02:45:04.583957 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63b375e9-d94c-4e34-882c-be534f9c046b" path="/var/lib/kubelet/pods/63b375e9-d94c-4e34-882c-be534f9c046b/volumes" Dec 03 02:45:14 crc kubenswrapper[4912]: I1203 02:45:14.578729 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:45:14 crc kubenswrapper[4912]: E1203 02:45:14.579380 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:45:28 crc kubenswrapper[4912]: I1203 02:45:28.572397 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:45:28 crc kubenswrapper[4912]: E1203 02:45:28.573168 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:45:30 crc kubenswrapper[4912]: I1203 02:45:30.790961 4912 scope.go:117] "RemoveContainer" containerID="d0cb19268c64db4fbee75405ec47954865f2fc9600efd9186bf7a98283a0172a" Dec 03 02:45:39 crc kubenswrapper[4912]: I1203 02:45:39.021633 4912 generic.go:334] "Generic (PLEG): container finished" podID="b7f11bf1-cada-4639-b216-dbd785813d62" containerID="7f43699422378b073dcf66075391c2e53b1c36d621a513cb4192f1e6e964dd04" exitCode=0 Dec 03 02:45:39 crc kubenswrapper[4912]: I1203 02:45:39.021697 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/crc-debug-6rmxg" event={"ID":"b7f11bf1-cada-4639-b216-dbd785813d62","Type":"ContainerDied","Data":"7f43699422378b073dcf66075391c2e53b1c36d621a513cb4192f1e6e964dd04"} Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.150356 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.173624 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7f11bf1-cada-4639-b216-dbd785813d62-host\") pod \"b7f11bf1-cada-4639-b216-dbd785813d62\" (UID: \"b7f11bf1-cada-4639-b216-dbd785813d62\") " Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.174010 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgmpd\" (UniqueName: \"kubernetes.io/projected/b7f11bf1-cada-4639-b216-dbd785813d62-kube-api-access-kgmpd\") pod \"b7f11bf1-cada-4639-b216-dbd785813d62\" (UID: \"b7f11bf1-cada-4639-b216-dbd785813d62\") " Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.175575 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7f11bf1-cada-4639-b216-dbd785813d62-host" (OuterVolumeSpecName: "host") pod "b7f11bf1-cada-4639-b216-dbd785813d62" (UID: "b7f11bf1-cada-4639-b216-dbd785813d62"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.183203 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f11bf1-cada-4639-b216-dbd785813d62-kube-api-access-kgmpd" (OuterVolumeSpecName: "kube-api-access-kgmpd") pod "b7f11bf1-cada-4639-b216-dbd785813d62" (UID: "b7f11bf1-cada-4639-b216-dbd785813d62"). InnerVolumeSpecName "kube-api-access-kgmpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.193158 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x4djm/crc-debug-6rmxg"] Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.206661 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x4djm/crc-debug-6rmxg"] Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.276731 4912 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b7f11bf1-cada-4639-b216-dbd785813d62-host\") on node \"crc\" DevicePath \"\"" Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.276775 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgmpd\" (UniqueName: \"kubernetes.io/projected/b7f11bf1-cada-4639-b216-dbd785813d62-kube-api-access-kgmpd\") on node \"crc\" DevicePath \"\"" Dec 03 02:45:40 crc kubenswrapper[4912]: I1203 02:45:40.597338 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7f11bf1-cada-4639-b216-dbd785813d62" path="/var/lib/kubelet/pods/b7f11bf1-cada-4639-b216-dbd785813d62/volumes" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.045488 4912 scope.go:117] "RemoveContainer" containerID="7f43699422378b073dcf66075391c2e53b1c36d621a513cb4192f1e6e964dd04" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.045562 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-6rmxg" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.442184 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x4djm/crc-debug-nfh9z"] Dec 03 02:45:41 crc kubenswrapper[4912]: E1203 02:45:41.442814 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d14eb280-2992-4a7c-96b5-64e4e8d75c6f" containerName="collect-profiles" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.442830 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d14eb280-2992-4a7c-96b5-64e4e8d75c6f" containerName="collect-profiles" Dec 03 02:45:41 crc kubenswrapper[4912]: E1203 02:45:41.442877 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f11bf1-cada-4639-b216-dbd785813d62" containerName="container-00" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.442885 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f11bf1-cada-4639-b216-dbd785813d62" containerName="container-00" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.443177 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7f11bf1-cada-4639-b216-dbd785813d62" containerName="container-00" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.443196 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d14eb280-2992-4a7c-96b5-64e4e8d75c6f" containerName="collect-profiles" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.444115 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.447265 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x4djm"/"default-dockercfg-fhkb2" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.504277 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqz6h\" (UniqueName: \"kubernetes.io/projected/d0fa4064-c32d-4991-8b68-5ea920f22ed7-kube-api-access-kqz6h\") pod \"crc-debug-nfh9z\" (UID: \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\") " pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.504698 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d0fa4064-c32d-4991-8b68-5ea920f22ed7-host\") pod \"crc-debug-nfh9z\" (UID: \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\") " pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.571923 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:45:41 crc kubenswrapper[4912]: E1203 02:45:41.572274 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.606722 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqz6h\" (UniqueName: \"kubernetes.io/projected/d0fa4064-c32d-4991-8b68-5ea920f22ed7-kube-api-access-kqz6h\") pod \"crc-debug-nfh9z\" (UID: \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\") " pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.607092 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d0fa4064-c32d-4991-8b68-5ea920f22ed7-host\") pod \"crc-debug-nfh9z\" (UID: \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\") " pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.607348 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d0fa4064-c32d-4991-8b68-5ea920f22ed7-host\") pod \"crc-debug-nfh9z\" (UID: \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\") " pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.625971 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqz6h\" (UniqueName: \"kubernetes.io/projected/d0fa4064-c32d-4991-8b68-5ea920f22ed7-kube-api-access-kqz6h\") pod \"crc-debug-nfh9z\" (UID: \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\") " pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:41 crc kubenswrapper[4912]: I1203 02:45:41.764748 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:42 crc kubenswrapper[4912]: I1203 02:45:42.058348 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/crc-debug-nfh9z" event={"ID":"d0fa4064-c32d-4991-8b68-5ea920f22ed7","Type":"ContainerStarted","Data":"69f95eb5a5c4b8f9990d32a101cbc5fabeaf5de6e1b3941df524af5c7a7c8d73"} Dec 03 02:45:43 crc kubenswrapper[4912]: I1203 02:45:43.069371 4912 generic.go:334] "Generic (PLEG): container finished" podID="d0fa4064-c32d-4991-8b68-5ea920f22ed7" containerID="b6a117927523a3ecf9be155eb08cf11ab9a062c34795de15506c089c9e072cbe" exitCode=0 Dec 03 02:45:43 crc kubenswrapper[4912]: I1203 02:45:43.069472 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/crc-debug-nfh9z" event={"ID":"d0fa4064-c32d-4991-8b68-5ea920f22ed7","Type":"ContainerDied","Data":"b6a117927523a3ecf9be155eb08cf11ab9a062c34795de15506c089c9e072cbe"} Dec 03 02:45:44 crc kubenswrapper[4912]: I1203 02:45:44.188604 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:44 crc kubenswrapper[4912]: I1203 02:45:44.274596 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqz6h\" (UniqueName: \"kubernetes.io/projected/d0fa4064-c32d-4991-8b68-5ea920f22ed7-kube-api-access-kqz6h\") pod \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\" (UID: \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\") " Dec 03 02:45:44 crc kubenswrapper[4912]: I1203 02:45:44.274736 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d0fa4064-c32d-4991-8b68-5ea920f22ed7-host\") pod \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\" (UID: \"d0fa4064-c32d-4991-8b68-5ea920f22ed7\") " Dec 03 02:45:44 crc kubenswrapper[4912]: I1203 02:45:44.274837 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0fa4064-c32d-4991-8b68-5ea920f22ed7-host" (OuterVolumeSpecName: "host") pod "d0fa4064-c32d-4991-8b68-5ea920f22ed7" (UID: "d0fa4064-c32d-4991-8b68-5ea920f22ed7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 02:45:44 crc kubenswrapper[4912]: I1203 02:45:44.275516 4912 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d0fa4064-c32d-4991-8b68-5ea920f22ed7-host\") on node \"crc\" DevicePath \"\"" Dec 03 02:45:44 crc kubenswrapper[4912]: I1203 02:45:44.292232 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0fa4064-c32d-4991-8b68-5ea920f22ed7-kube-api-access-kqz6h" (OuterVolumeSpecName: "kube-api-access-kqz6h") pod "d0fa4064-c32d-4991-8b68-5ea920f22ed7" (UID: "d0fa4064-c32d-4991-8b68-5ea920f22ed7"). InnerVolumeSpecName "kube-api-access-kqz6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:45:44 crc kubenswrapper[4912]: I1203 02:45:44.377407 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqz6h\" (UniqueName: \"kubernetes.io/projected/d0fa4064-c32d-4991-8b68-5ea920f22ed7-kube-api-access-kqz6h\") on node \"crc\" DevicePath \"\"" Dec 03 02:45:45 crc kubenswrapper[4912]: I1203 02:45:45.089501 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/crc-debug-nfh9z" event={"ID":"d0fa4064-c32d-4991-8b68-5ea920f22ed7","Type":"ContainerDied","Data":"69f95eb5a5c4b8f9990d32a101cbc5fabeaf5de6e1b3941df524af5c7a7c8d73"} Dec 03 02:45:45 crc kubenswrapper[4912]: I1203 02:45:45.089748 4912 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69f95eb5a5c4b8f9990d32a101cbc5fabeaf5de6e1b3941df524af5c7a7c8d73" Dec 03 02:45:45 crc kubenswrapper[4912]: I1203 02:45:45.089575 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-nfh9z" Dec 03 02:45:45 crc kubenswrapper[4912]: I1203 02:45:45.445997 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x4djm/crc-debug-nfh9z"] Dec 03 02:45:45 crc kubenswrapper[4912]: I1203 02:45:45.457209 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x4djm/crc-debug-nfh9z"] Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.592871 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0fa4064-c32d-4991-8b68-5ea920f22ed7" path="/var/lib/kubelet/pods/d0fa4064-c32d-4991-8b68-5ea920f22ed7/volumes" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.669223 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-x4djm/crc-debug-mz67w"] Dec 03 02:45:46 crc kubenswrapper[4912]: E1203 02:45:46.670981 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fa4064-c32d-4991-8b68-5ea920f22ed7" containerName="container-00" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.671013 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fa4064-c32d-4991-8b68-5ea920f22ed7" containerName="container-00" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.671291 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0fa4064-c32d-4991-8b68-5ea920f22ed7" containerName="container-00" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.672310 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.677253 4912 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-x4djm"/"default-dockercfg-fhkb2" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.842558 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xrwj\" (UniqueName: \"kubernetes.io/projected/0d6e2b43-dc3e-4951-8b35-fdb94904993d-kube-api-access-5xrwj\") pod \"crc-debug-mz67w\" (UID: \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\") " pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.842636 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d6e2b43-dc3e-4951-8b35-fdb94904993d-host\") pod \"crc-debug-mz67w\" (UID: \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\") " pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.946017 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xrwj\" (UniqueName: \"kubernetes.io/projected/0d6e2b43-dc3e-4951-8b35-fdb94904993d-kube-api-access-5xrwj\") pod \"crc-debug-mz67w\" (UID: \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\") " pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.946141 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d6e2b43-dc3e-4951-8b35-fdb94904993d-host\") pod \"crc-debug-mz67w\" (UID: \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\") " pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.946298 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d6e2b43-dc3e-4951-8b35-fdb94904993d-host\") pod \"crc-debug-mz67w\" (UID: \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\") " pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:46 crc kubenswrapper[4912]: I1203 02:45:46.975566 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xrwj\" (UniqueName: \"kubernetes.io/projected/0d6e2b43-dc3e-4951-8b35-fdb94904993d-kube-api-access-5xrwj\") pod \"crc-debug-mz67w\" (UID: \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\") " pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:47 crc kubenswrapper[4912]: I1203 02:45:47.003999 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:47 crc kubenswrapper[4912]: I1203 02:45:47.121851 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/crc-debug-mz67w" event={"ID":"0d6e2b43-dc3e-4951-8b35-fdb94904993d","Type":"ContainerStarted","Data":"3f8888308f1454779891bba796b59554610e3cbaa9c296765e15458c39f4b1f0"} Dec 03 02:45:48 crc kubenswrapper[4912]: I1203 02:45:48.143835 4912 generic.go:334] "Generic (PLEG): container finished" podID="0d6e2b43-dc3e-4951-8b35-fdb94904993d" containerID="1341baf68e031035bc847a6032eeffdb956a1dc82c0e45f02e059f0fbd955a78" exitCode=0 Dec 03 02:45:48 crc kubenswrapper[4912]: I1203 02:45:48.144488 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/crc-debug-mz67w" event={"ID":"0d6e2b43-dc3e-4951-8b35-fdb94904993d","Type":"ContainerDied","Data":"1341baf68e031035bc847a6032eeffdb956a1dc82c0e45f02e059f0fbd955a78"} Dec 03 02:45:48 crc kubenswrapper[4912]: I1203 02:45:48.205893 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x4djm/crc-debug-mz67w"] Dec 03 02:45:48 crc kubenswrapper[4912]: I1203 02:45:48.220166 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x4djm/crc-debug-mz67w"] Dec 03 02:45:49 crc kubenswrapper[4912]: I1203 02:45:49.272142 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:49 crc kubenswrapper[4912]: I1203 02:45:49.409221 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xrwj\" (UniqueName: \"kubernetes.io/projected/0d6e2b43-dc3e-4951-8b35-fdb94904993d-kube-api-access-5xrwj\") pod \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\" (UID: \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\") " Dec 03 02:45:49 crc kubenswrapper[4912]: I1203 02:45:49.409324 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d6e2b43-dc3e-4951-8b35-fdb94904993d-host\") pod \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\" (UID: \"0d6e2b43-dc3e-4951-8b35-fdb94904993d\") " Dec 03 02:45:49 crc kubenswrapper[4912]: I1203 02:45:49.409452 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d6e2b43-dc3e-4951-8b35-fdb94904993d-host" (OuterVolumeSpecName: "host") pod "0d6e2b43-dc3e-4951-8b35-fdb94904993d" (UID: "0d6e2b43-dc3e-4951-8b35-fdb94904993d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 02:45:49 crc kubenswrapper[4912]: I1203 02:45:49.409956 4912 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d6e2b43-dc3e-4951-8b35-fdb94904993d-host\") on node \"crc\" DevicePath \"\"" Dec 03 02:45:49 crc kubenswrapper[4912]: I1203 02:45:49.424737 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d6e2b43-dc3e-4951-8b35-fdb94904993d-kube-api-access-5xrwj" (OuterVolumeSpecName: "kube-api-access-5xrwj") pod "0d6e2b43-dc3e-4951-8b35-fdb94904993d" (UID: "0d6e2b43-dc3e-4951-8b35-fdb94904993d"). InnerVolumeSpecName "kube-api-access-5xrwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:45:49 crc kubenswrapper[4912]: I1203 02:45:49.512315 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xrwj\" (UniqueName: \"kubernetes.io/projected/0d6e2b43-dc3e-4951-8b35-fdb94904993d-kube-api-access-5xrwj\") on node \"crc\" DevicePath \"\"" Dec 03 02:45:50 crc kubenswrapper[4912]: I1203 02:45:50.164186 4912 scope.go:117] "RemoveContainer" containerID="1341baf68e031035bc847a6032eeffdb956a1dc82c0e45f02e059f0fbd955a78" Dec 03 02:45:50 crc kubenswrapper[4912]: I1203 02:45:50.164208 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/crc-debug-mz67w" Dec 03 02:45:50 crc kubenswrapper[4912]: I1203 02:45:50.583991 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d6e2b43-dc3e-4951-8b35-fdb94904993d" path="/var/lib/kubelet/pods/0d6e2b43-dc3e-4951-8b35-fdb94904993d/volumes" Dec 03 02:45:55 crc kubenswrapper[4912]: I1203 02:45:55.572621 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:45:55 crc kubenswrapper[4912]: E1203 02:45:55.573190 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:46:09 crc kubenswrapper[4912]: I1203 02:46:09.572548 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:46:09 crc kubenswrapper[4912]: E1203 02:46:09.573851 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:46:24 crc kubenswrapper[4912]: I1203 02:46:24.590063 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:46:24 crc kubenswrapper[4912]: E1203 02:46:24.591245 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:46:37 crc kubenswrapper[4912]: I1203 02:46:37.571775 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:46:37 crc kubenswrapper[4912]: E1203 02:46:37.572539 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.040556 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xqbh7"] Dec 03 02:46:43 crc kubenswrapper[4912]: E1203 02:46:43.042291 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d6e2b43-dc3e-4951-8b35-fdb94904993d" containerName="container-00" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.042305 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d6e2b43-dc3e-4951-8b35-fdb94904993d" containerName="container-00" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.042579 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d6e2b43-dc3e-4951-8b35-fdb94904993d" containerName="container-00" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.044339 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.071212 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xqbh7"] Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.171387 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-utilities\") pod \"community-operators-xqbh7\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.171516 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g47k\" (UniqueName: \"kubernetes.io/projected/35140842-fcec-49dc-a3a2-3ecfad78e611-kube-api-access-4g47k\") pod \"community-operators-xqbh7\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.171721 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-catalog-content\") pod \"community-operators-xqbh7\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.273363 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-catalog-content\") pod \"community-operators-xqbh7\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.273457 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-utilities\") pod \"community-operators-xqbh7\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.273552 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g47k\" (UniqueName: \"kubernetes.io/projected/35140842-fcec-49dc-a3a2-3ecfad78e611-kube-api-access-4g47k\") pod \"community-operators-xqbh7\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.274186 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-catalog-content\") pod \"community-operators-xqbh7\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.274410 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-utilities\") pod \"community-operators-xqbh7\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.295413 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g47k\" (UniqueName: \"kubernetes.io/projected/35140842-fcec-49dc-a3a2-3ecfad78e611-kube-api-access-4g47k\") pod \"community-operators-xqbh7\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:43 crc kubenswrapper[4912]: I1203 02:46:43.363866 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:44 crc kubenswrapper[4912]: I1203 02:46:44.013976 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xqbh7"] Dec 03 02:46:44 crc kubenswrapper[4912]: I1203 02:46:44.820942 4912 generic.go:334] "Generic (PLEG): container finished" podID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerID="dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8" exitCode=0 Dec 03 02:46:44 crc kubenswrapper[4912]: I1203 02:46:44.821735 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqbh7" event={"ID":"35140842-fcec-49dc-a3a2-3ecfad78e611","Type":"ContainerDied","Data":"dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8"} Dec 03 02:46:44 crc kubenswrapper[4912]: I1203 02:46:44.821825 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqbh7" event={"ID":"35140842-fcec-49dc-a3a2-3ecfad78e611","Type":"ContainerStarted","Data":"6463554afcd2beb1c1ce22147ec7d7eccc189e4921b8f5df4df66f5e6ab77184"} Dec 03 02:46:44 crc kubenswrapper[4912]: I1203 02:46:44.823047 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 02:46:45 crc kubenswrapper[4912]: I1203 02:46:45.854103 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g2fm8"] Dec 03 02:46:45 crc kubenswrapper[4912]: I1203 02:46:45.856621 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:45 crc kubenswrapper[4912]: I1203 02:46:45.864209 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g2fm8"] Dec 03 02:46:45 crc kubenswrapper[4912]: I1203 02:46:45.967138 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-catalog-content\") pod \"redhat-marketplace-g2fm8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:45 crc kubenswrapper[4912]: I1203 02:46:45.967765 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sfwn\" (UniqueName: \"kubernetes.io/projected/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-kube-api-access-5sfwn\") pod \"redhat-marketplace-g2fm8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:45 crc kubenswrapper[4912]: I1203 02:46:45.968637 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-utilities\") pod \"redhat-marketplace-g2fm8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:46 crc kubenswrapper[4912]: I1203 02:46:46.070602 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-catalog-content\") pod \"redhat-marketplace-g2fm8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:46 crc kubenswrapper[4912]: I1203 02:46:46.070665 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sfwn\" (UniqueName: \"kubernetes.io/projected/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-kube-api-access-5sfwn\") pod \"redhat-marketplace-g2fm8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:46 crc kubenswrapper[4912]: I1203 02:46:46.070816 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-utilities\") pod \"redhat-marketplace-g2fm8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:46 crc kubenswrapper[4912]: I1203 02:46:46.071112 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-catalog-content\") pod \"redhat-marketplace-g2fm8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:46 crc kubenswrapper[4912]: I1203 02:46:46.071151 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-utilities\") pod \"redhat-marketplace-g2fm8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:46 crc kubenswrapper[4912]: I1203 02:46:46.098392 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sfwn\" (UniqueName: \"kubernetes.io/projected/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-kube-api-access-5sfwn\") pod \"redhat-marketplace-g2fm8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:46 crc kubenswrapper[4912]: I1203 02:46:46.225123 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:46 crc kubenswrapper[4912]: I1203 02:46:46.852975 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqbh7" event={"ID":"35140842-fcec-49dc-a3a2-3ecfad78e611","Type":"ContainerStarted","Data":"c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109"} Dec 03 02:46:46 crc kubenswrapper[4912]: I1203 02:46:46.906915 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g2fm8"] Dec 03 02:46:46 crc kubenswrapper[4912]: W1203 02:46:46.976383 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b8caf38_5e5c_4f03_976b_45d693b3dcf8.slice/crio-05106b15db5d50b7d3a2d7b5342f47540589f98626ec7b5c5ced43d85db92b41 WatchSource:0}: Error finding container 05106b15db5d50b7d3a2d7b5342f47540589f98626ec7b5c5ced43d85db92b41: Status 404 returned error can't find the container with id 05106b15db5d50b7d3a2d7b5342f47540589f98626ec7b5c5ced43d85db92b41 Dec 03 02:46:47 crc kubenswrapper[4912]: I1203 02:46:47.864231 4912 generic.go:334] "Generic (PLEG): container finished" podID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerID="1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba" exitCode=0 Dec 03 02:46:47 crc kubenswrapper[4912]: I1203 02:46:47.864551 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2fm8" event={"ID":"7b8caf38-5e5c-4f03-976b-45d693b3dcf8","Type":"ContainerDied","Data":"1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba"} Dec 03 02:46:47 crc kubenswrapper[4912]: I1203 02:46:47.864588 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2fm8" event={"ID":"7b8caf38-5e5c-4f03-976b-45d693b3dcf8","Type":"ContainerStarted","Data":"05106b15db5d50b7d3a2d7b5342f47540589f98626ec7b5c5ced43d85db92b41"} Dec 03 02:46:47 crc kubenswrapper[4912]: I1203 02:46:47.868473 4912 generic.go:334] "Generic (PLEG): container finished" podID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerID="c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109" exitCode=0 Dec 03 02:46:47 crc kubenswrapper[4912]: I1203 02:46:47.868502 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqbh7" event={"ID":"35140842-fcec-49dc-a3a2-3ecfad78e611","Type":"ContainerDied","Data":"c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109"} Dec 03 02:46:48 crc kubenswrapper[4912]: I1203 02:46:48.571837 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:46:48 crc kubenswrapper[4912]: E1203 02:46:48.572587 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:46:48 crc kubenswrapper[4912]: I1203 02:46:48.882008 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2fm8" event={"ID":"7b8caf38-5e5c-4f03-976b-45d693b3dcf8","Type":"ContainerStarted","Data":"84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba"} Dec 03 02:46:48 crc kubenswrapper[4912]: I1203 02:46:48.885187 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqbh7" event={"ID":"35140842-fcec-49dc-a3a2-3ecfad78e611","Type":"ContainerStarted","Data":"3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60"} Dec 03 02:46:48 crc kubenswrapper[4912]: I1203 02:46:48.932352 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xqbh7" podStartSLOduration=2.227831692 podStartE2EDuration="5.932327809s" podCreationTimestamp="2025-12-03 02:46:43 +0000 UTC" firstStartedPulling="2025-12-03 02:46:44.822851012 +0000 UTC m=+8590.464871572" lastFinishedPulling="2025-12-03 02:46:48.527347129 +0000 UTC m=+8594.169367689" observedRunningTime="2025-12-03 02:46:48.917913602 +0000 UTC m=+8594.559934182" watchObservedRunningTime="2025-12-03 02:46:48.932327809 +0000 UTC m=+8594.574348379" Dec 03 02:46:49 crc kubenswrapper[4912]: I1203 02:46:49.895277 4912 generic.go:334] "Generic (PLEG): container finished" podID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerID="84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba" exitCode=0 Dec 03 02:46:49 crc kubenswrapper[4912]: I1203 02:46:49.897447 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2fm8" event={"ID":"7b8caf38-5e5c-4f03-976b-45d693b3dcf8","Type":"ContainerDied","Data":"84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba"} Dec 03 02:46:50 crc kubenswrapper[4912]: I1203 02:46:50.910935 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2fm8" event={"ID":"7b8caf38-5e5c-4f03-976b-45d693b3dcf8","Type":"ContainerStarted","Data":"19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70"} Dec 03 02:46:50 crc kubenswrapper[4912]: I1203 02:46:50.933831 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g2fm8" podStartSLOduration=3.442776311 podStartE2EDuration="5.933812304s" podCreationTimestamp="2025-12-03 02:46:45 +0000 UTC" firstStartedPulling="2025-12-03 02:46:47.867946949 +0000 UTC m=+8593.509967509" lastFinishedPulling="2025-12-03 02:46:50.358982922 +0000 UTC m=+8596.001003502" observedRunningTime="2025-12-03 02:46:50.928849511 +0000 UTC m=+8596.570870071" watchObservedRunningTime="2025-12-03 02:46:50.933812304 +0000 UTC m=+8596.575832864" Dec 03 02:46:53 crc kubenswrapper[4912]: I1203 02:46:53.364223 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:53 crc kubenswrapper[4912]: I1203 02:46:53.364892 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:53 crc kubenswrapper[4912]: I1203 02:46:53.455327 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:54 crc kubenswrapper[4912]: I1203 02:46:54.033241 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:54 crc kubenswrapper[4912]: I1203 02:46:54.435379 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xqbh7"] Dec 03 02:46:55 crc kubenswrapper[4912]: I1203 02:46:55.976199 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xqbh7" podUID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerName="registry-server" containerID="cri-o://3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60" gracePeriod=2 Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.226139 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.235392 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.309393 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.661379 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.787112 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g47k\" (UniqueName: \"kubernetes.io/projected/35140842-fcec-49dc-a3a2-3ecfad78e611-kube-api-access-4g47k\") pod \"35140842-fcec-49dc-a3a2-3ecfad78e611\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.787177 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-utilities\") pod \"35140842-fcec-49dc-a3a2-3ecfad78e611\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.787203 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-catalog-content\") pod \"35140842-fcec-49dc-a3a2-3ecfad78e611\" (UID: \"35140842-fcec-49dc-a3a2-3ecfad78e611\") " Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.793459 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-utilities" (OuterVolumeSpecName: "utilities") pod "35140842-fcec-49dc-a3a2-3ecfad78e611" (UID: "35140842-fcec-49dc-a3a2-3ecfad78e611"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.795223 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35140842-fcec-49dc-a3a2-3ecfad78e611-kube-api-access-4g47k" (OuterVolumeSpecName: "kube-api-access-4g47k") pod "35140842-fcec-49dc-a3a2-3ecfad78e611" (UID: "35140842-fcec-49dc-a3a2-3ecfad78e611"). InnerVolumeSpecName "kube-api-access-4g47k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.852782 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35140842-fcec-49dc-a3a2-3ecfad78e611" (UID: "35140842-fcec-49dc-a3a2-3ecfad78e611"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.890496 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g47k\" (UniqueName: \"kubernetes.io/projected/35140842-fcec-49dc-a3a2-3ecfad78e611-kube-api-access-4g47k\") on node \"crc\" DevicePath \"\"" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.890536 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.890549 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35140842-fcec-49dc-a3a2-3ecfad78e611-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.993855 4912 generic.go:334] "Generic (PLEG): container finished" podID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerID="3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60" exitCode=0 Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.993916 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqbh7" event={"ID":"35140842-fcec-49dc-a3a2-3ecfad78e611","Type":"ContainerDied","Data":"3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60"} Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.994013 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqbh7" event={"ID":"35140842-fcec-49dc-a3a2-3ecfad78e611","Type":"ContainerDied","Data":"6463554afcd2beb1c1ce22147ec7d7eccc189e4921b8f5df4df66f5e6ab77184"} Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.993941 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqbh7" Dec 03 02:46:56 crc kubenswrapper[4912]: I1203 02:46:56.994056 4912 scope.go:117] "RemoveContainer" containerID="3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60" Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.051814 4912 scope.go:117] "RemoveContainer" containerID="c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109" Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.060341 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xqbh7"] Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.077069 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xqbh7"] Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.092874 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.094672 4912 scope.go:117] "RemoveContainer" containerID="dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8" Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.157213 4912 scope.go:117] "RemoveContainer" containerID="3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60" Dec 03 02:46:57 crc kubenswrapper[4912]: E1203 02:46:57.157902 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60\": container with ID starting with 3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60 not found: ID does not exist" containerID="3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60" Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.157961 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60"} err="failed to get container status \"3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60\": rpc error: code = NotFound desc = could not find container \"3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60\": container with ID starting with 3a5900f39958bd0cc6218181fb5b960e9e999928b65671d263e911fa47cc9a60 not found: ID does not exist" Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.157996 4912 scope.go:117] "RemoveContainer" containerID="c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109" Dec 03 02:46:57 crc kubenswrapper[4912]: E1203 02:46:57.158504 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109\": container with ID starting with c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109 not found: ID does not exist" containerID="c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109" Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.158549 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109"} err="failed to get container status \"c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109\": rpc error: code = NotFound desc = could not find container \"c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109\": container with ID starting with c25d2eb82cabb3eadeb41a156f96a3ff8c80045c69dc48e3d5296b03ffd3e109 not found: ID does not exist" Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.158576 4912 scope.go:117] "RemoveContainer" containerID="dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8" Dec 03 02:46:57 crc kubenswrapper[4912]: E1203 02:46:57.158962 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8\": container with ID starting with dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8 not found: ID does not exist" containerID="dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8" Dec 03 02:46:57 crc kubenswrapper[4912]: I1203 02:46:57.158996 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8"} err="failed to get container status \"dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8\": rpc error: code = NotFound desc = could not find container \"dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8\": container with ID starting with dbce707d86e9afdac45368140a70e69ef9bfad09663a3675c19bcf4db3828da8 not found: ID does not exist" Dec 03 02:46:58 crc kubenswrapper[4912]: I1203 02:46:58.599573 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35140842-fcec-49dc-a3a2-3ecfad78e611" path="/var/lib/kubelet/pods/35140842-fcec-49dc-a3a2-3ecfad78e611/volumes" Dec 03 02:46:59 crc kubenswrapper[4912]: I1203 02:46:59.435144 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g2fm8"] Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.038244 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g2fm8" podUID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerName="registry-server" containerID="cri-o://19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70" gracePeriod=2 Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.553896 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.652123 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-catalog-content\") pod \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.652265 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sfwn\" (UniqueName: \"kubernetes.io/projected/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-kube-api-access-5sfwn\") pod \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.652330 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-utilities\") pod \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\" (UID: \"7b8caf38-5e5c-4f03-976b-45d693b3dcf8\") " Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.653262 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-utilities" (OuterVolumeSpecName: "utilities") pod "7b8caf38-5e5c-4f03-976b-45d693b3dcf8" (UID: "7b8caf38-5e5c-4f03-976b-45d693b3dcf8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.653957 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.662245 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-kube-api-access-5sfwn" (OuterVolumeSpecName: "kube-api-access-5sfwn") pod "7b8caf38-5e5c-4f03-976b-45d693b3dcf8" (UID: "7b8caf38-5e5c-4f03-976b-45d693b3dcf8"). InnerVolumeSpecName "kube-api-access-5sfwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.684535 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b8caf38-5e5c-4f03-976b-45d693b3dcf8" (UID: "7b8caf38-5e5c-4f03-976b-45d693b3dcf8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.755840 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:47:00 crc kubenswrapper[4912]: I1203 02:47:00.755876 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sfwn\" (UniqueName: \"kubernetes.io/projected/7b8caf38-5e5c-4f03-976b-45d693b3dcf8-kube-api-access-5sfwn\") on node \"crc\" DevicePath \"\"" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.050068 4912 generic.go:334] "Generic (PLEG): container finished" podID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerID="19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70" exitCode=0 Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.050113 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2fm8" event={"ID":"7b8caf38-5e5c-4f03-976b-45d693b3dcf8","Type":"ContainerDied","Data":"19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70"} Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.050141 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g2fm8" event={"ID":"7b8caf38-5e5c-4f03-976b-45d693b3dcf8","Type":"ContainerDied","Data":"05106b15db5d50b7d3a2d7b5342f47540589f98626ec7b5c5ced43d85db92b41"} Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.050139 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g2fm8" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.050159 4912 scope.go:117] "RemoveContainer" containerID="19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.080609 4912 scope.go:117] "RemoveContainer" containerID="84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.085912 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g2fm8"] Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.146197 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g2fm8"] Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.159105 4912 scope.go:117] "RemoveContainer" containerID="1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.196664 4912 scope.go:117] "RemoveContainer" containerID="19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70" Dec 03 02:47:01 crc kubenswrapper[4912]: E1203 02:47:01.197241 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70\": container with ID starting with 19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70 not found: ID does not exist" containerID="19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.197284 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70"} err="failed to get container status \"19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70\": rpc error: code = NotFound desc = could not find container \"19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70\": container with ID starting with 19bcade68ccd6415dcc7441a313e1f37d10629e3dc41068c72fad7ea1f4d5f70 not found: ID does not exist" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.197312 4912 scope.go:117] "RemoveContainer" containerID="84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba" Dec 03 02:47:01 crc kubenswrapper[4912]: E1203 02:47:01.197795 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba\": container with ID starting with 84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba not found: ID does not exist" containerID="84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.197835 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba"} err="failed to get container status \"84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba\": rpc error: code = NotFound desc = could not find container \"84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba\": container with ID starting with 84733c58ae2cbf49d89a7863db635ccf88f02295aa130732142e966b6691e1ba not found: ID does not exist" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.197876 4912 scope.go:117] "RemoveContainer" containerID="1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba" Dec 03 02:47:01 crc kubenswrapper[4912]: E1203 02:47:01.198252 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba\": container with ID starting with 1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba not found: ID does not exist" containerID="1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.198276 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba"} err="failed to get container status \"1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba\": rpc error: code = NotFound desc = could not find container \"1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba\": container with ID starting with 1944348ace321434c1428405288c0f449d4c750a8231337e6e848a15ec35e4ba not found: ID does not exist" Dec 03 02:47:01 crc kubenswrapper[4912]: I1203 02:47:01.571381 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:47:01 crc kubenswrapper[4912]: E1203 02:47:01.571754 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:47:02 crc kubenswrapper[4912]: I1203 02:47:02.537477 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_41945ca7-6d5e-4875-b770-801813413434/aodh-api/0.log" Dec 03 02:47:02 crc kubenswrapper[4912]: I1203 02:47:02.603272 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" path="/var/lib/kubelet/pods/7b8caf38-5e5c-4f03-976b-45d693b3dcf8/volumes" Dec 03 02:47:02 crc kubenswrapper[4912]: I1203 02:47:02.765085 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_41945ca7-6d5e-4875-b770-801813413434/aodh-evaluator/0.log" Dec 03 02:47:02 crc kubenswrapper[4912]: I1203 02:47:02.811655 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_41945ca7-6d5e-4875-b770-801813413434/aodh-listener/0.log" Dec 03 02:47:02 crc kubenswrapper[4912]: I1203 02:47:02.819194 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_41945ca7-6d5e-4875-b770-801813413434/aodh-notifier/0.log" Dec 03 02:47:02 crc kubenswrapper[4912]: I1203 02:47:02.991934 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-75f549cc78-m5rcp_21e2b86e-eeab-4af6-b68d-8f67abda11e2/barbican-api/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.009395 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-75f549cc78-m5rcp_21e2b86e-eeab-4af6-b68d-8f67abda11e2/barbican-api-log/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.161978 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6f4c446894-wlqxw_3269ec0e-5405-4a8d-b7ea-e07c5a7a7212/barbican-keystone-listener/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.300177 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-79459d7cd7-qdq5s_c6d0b29e-b4cd-4f78-b9f7-200691c64f93/barbican-worker/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.365914 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-79459d7cd7-qdq5s_c6d0b29e-b4cd-4f78-b9f7-200691c64f93/barbican-worker-log/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.400919 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6f4c446894-wlqxw_3269ec0e-5405-4a8d-b7ea-e07c5a7a7212/barbican-keystone-listener-log/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.580083 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-qqkdr_ac32e469-f344-4ee1-8702-bfa8b124ac35/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.726887 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bbc8b9dd-35e3-4d6b-8df6-e44368688196/ceilometer-central-agent/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.779770 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bbc8b9dd-35e3-4d6b-8df6-e44368688196/ceilometer-notification-agent/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.820926 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bbc8b9dd-35e3-4d6b-8df6-e44368688196/proxy-httpd/0.log" Dec 03 02:47:03 crc kubenswrapper[4912]: I1203 02:47:03.851190 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_bbc8b9dd-35e3-4d6b-8df6-e44368688196/sg-core/0.log" Dec 03 02:47:04 crc kubenswrapper[4912]: I1203 02:47:04.026017 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-ffjtv_b529c480-6fb1-4a0b-a915-1aca2f6e85bd/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:04 crc kubenswrapper[4912]: I1203 02:47:04.087792 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-25hxg_d241940f-a3d4-4c91-93c7-2149a6e25ae2/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:04 crc kubenswrapper[4912]: I1203 02:47:04.387565 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_64841d14-78b5-4737-90ec-b812f2994397/cinder-api/0.log" Dec 03 02:47:04 crc kubenswrapper[4912]: I1203 02:47:04.432359 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_64841d14-78b5-4737-90ec-b812f2994397/cinder-api-log/0.log" Dec 03 02:47:04 crc kubenswrapper[4912]: I1203 02:47:04.713073 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7/probe/0.log" Dec 03 02:47:04 crc kubenswrapper[4912]: I1203 02:47:04.825338 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0/cinder-scheduler/0.log" Dec 03 02:47:04 crc kubenswrapper[4912]: I1203 02:47:04.894057 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_51a45baf-7e2e-4e15-8945-6c6a0f2e1dd7/cinder-backup/0.log" Dec 03 02:47:04 crc kubenswrapper[4912]: I1203 02:47:04.992394 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8af6e3fe-8a7a-480e-80b6-cd5d5e634fd0/probe/0.log" Dec 03 02:47:05 crc kubenswrapper[4912]: I1203 02:47:05.199217 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3e1c22b1-b578-4957-98b5-b981f1f7dee2/probe/0.log" Dec 03 02:47:05 crc kubenswrapper[4912]: I1203 02:47:05.216606 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_3e1c22b1-b578-4957-98b5-b981f1f7dee2/cinder-volume/0.log" Dec 03 02:47:05 crc kubenswrapper[4912]: I1203 02:47:05.327356 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-mv8vv_50268f51-4c6d-4837-8806-3f741ee7b48c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:05 crc kubenswrapper[4912]: I1203 02:47:05.428679 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-78l9p_12198d9f-b54f-428d-9a81-7d6d2d44f695/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:05 crc kubenswrapper[4912]: I1203 02:47:05.507893 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-p2gz6_51b6bb89-372c-4193-ab71-6317807c1820/init/0.log" Dec 03 02:47:05 crc kubenswrapper[4912]: I1203 02:47:05.751568 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-p2gz6_51b6bb89-372c-4193-ab71-6317807c1820/init/0.log" Dec 03 02:47:05 crc kubenswrapper[4912]: I1203 02:47:05.823412 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bf551edf-1401-42c2-8d75-6dec09a149c6/glance-httpd/0.log" Dec 03 02:47:05 crc kubenswrapper[4912]: I1203 02:47:05.852131 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-c8d8d886c-p2gz6_51b6bb89-372c-4193-ab71-6317807c1820/dnsmasq-dns/0.log" Dec 03 02:47:05 crc kubenswrapper[4912]: I1203 02:47:05.964213 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_bf551edf-1401-42c2-8d75-6dec09a149c6/glance-log/0.log" Dec 03 02:47:06 crc kubenswrapper[4912]: I1203 02:47:06.079628 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17260561-5216-45fa-b547-2cbe178094d1/glance-log/0.log" Dec 03 02:47:06 crc kubenswrapper[4912]: I1203 02:47:06.105086 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_17260561-5216-45fa-b547-2cbe178094d1/glance-httpd/0.log" Dec 03 02:47:06 crc kubenswrapper[4912]: I1203 02:47:06.521073 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-678985f55c-wftgz_e31eb41a-2842-45fe-ba2d-6a84a048723f/heat-engine/0.log" Dec 03 02:47:07 crc kubenswrapper[4912]: I1203 02:47:07.097866 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d4dbbf8d-5vgmn_3d92b6a2-0079-4d39-9c11-9a31c22429f4/horizon/0.log" Dec 03 02:47:07 crc kubenswrapper[4912]: I1203 02:47:07.377517 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-ssn98_d2d7b572-6166-47b3-9346-17f881d869a7/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:07 crc kubenswrapper[4912]: I1203 02:47:07.405289 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-56b4585b94-qjwmb_23ac5ef2-d307-4073-982f-afe777413b41/heat-api/0.log" Dec 03 02:47:07 crc kubenswrapper[4912]: I1203 02:47:07.545039 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-5cdb6985d5-xgpkb_b0e666e2-5db8-409a-9a47-0da8fe895f75/heat-cfnapi/0.log" Dec 03 02:47:07 crc kubenswrapper[4912]: I1203 02:47:07.571740 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-jckxr_90fce4e2-3ec4-469d-a50a-8c8df048c2ab/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:07 crc kubenswrapper[4912]: I1203 02:47:07.631875 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d4dbbf8d-5vgmn_3d92b6a2-0079-4d39-9c11-9a31c22429f4/horizon-log/0.log" Dec 03 02:47:07 crc kubenswrapper[4912]: I1203 02:47:07.662107 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412061-jvbbz_65a4a770-16d8-4ea3-ac4a-9810ed0c2477/keystone-cron/0.log" Dec 03 02:47:07 crc kubenswrapper[4912]: I1203 02:47:07.785964 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29412121-dkhrq_bb6b4855-3d83-4341-942e-de0bc1027bc4/keystone-cron/0.log" Dec 03 02:47:07 crc kubenswrapper[4912]: I1203 02:47:07.980146 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_4ef4169e-dcdf-4d6d-a8e7-ea75f1194cba/kube-state-metrics/0.log" Dec 03 02:47:08 crc kubenswrapper[4912]: I1203 02:47:08.138023 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-q9btl_911b9678-7127-4ae0-ba4b-c059dc13796f/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:08 crc kubenswrapper[4912]: I1203 02:47:08.300150 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-46v4t_77e171a1-53b6-4a5d-94a2-2d9b42e1e71f/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:08 crc kubenswrapper[4912]: I1203 02:47:08.368158 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-d9977975b-bq8m5_a4be9da8-9ead-47ff-9bf5-ba2eb9f7251b/keystone-api/0.log" Dec 03 02:47:08 crc kubenswrapper[4912]: I1203 02:47:08.528060 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_eedcff7d-7bb2-4c30-8166-e7a2f22ac60e/manila-api/0.log" Dec 03 02:47:08 crc kubenswrapper[4912]: I1203 02:47:08.550672 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_eedcff7d-7bb2-4c30-8166-e7a2f22ac60e/manila-api-log/0.log" Dec 03 02:47:08 crc kubenswrapper[4912]: I1203 02:47:08.725088 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_dc97be40-fb7b-4af2-af77-45f0f360c8cb/probe/0.log" Dec 03 02:47:08 crc kubenswrapper[4912]: I1203 02:47:08.816873 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_dc97be40-fb7b-4af2-af77-45f0f360c8cb/manila-scheduler/0.log" Dec 03 02:47:08 crc kubenswrapper[4912]: I1203 02:47:08.910035 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b2a55784-64e0-4397-a975-c0b069cbd08f/probe/0.log" Dec 03 02:47:08 crc kubenswrapper[4912]: I1203 02:47:08.935916 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b2a55784-64e0-4397-a975-c0b069cbd08f/manila-share/0.log" Dec 03 02:47:09 crc kubenswrapper[4912]: I1203 02:47:09.127894 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_610be762-7f1b-4435-8764-eeac647f901c/mysqld-exporter/0.log" Dec 03 02:47:09 crc kubenswrapper[4912]: I1203 02:47:09.476833 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59fd659dd5-vmcdc_6b4100ae-eb0e-402b-b442-d2d313e2bff4/neutron-api/0.log" Dec 03 02:47:09 crc kubenswrapper[4912]: I1203 02:47:09.493754 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-4m747_748f1896-9a69-406c-bc59-0662de1da921/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:09 crc kubenswrapper[4912]: I1203 02:47:09.643644 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-59fd659dd5-vmcdc_6b4100ae-eb0e-402b-b442-d2d313e2bff4/neutron-httpd/0.log" Dec 03 02:47:10 crc kubenswrapper[4912]: I1203 02:47:10.369124 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2d46d472-d813-4f6d-b58a-8d671244960a/nova-cell0-conductor-conductor/0.log" Dec 03 02:47:10 crc kubenswrapper[4912]: I1203 02:47:10.566327 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23c2e51b-07f1-4b86-bc02-29969db68f38/nova-api-log/0.log" Dec 03 02:47:10 crc kubenswrapper[4912]: I1203 02:47:10.587193 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_8116ee8d-aa67-4a8f-a765-b846f397f7d4/nova-cell1-conductor-conductor/0.log" Dec 03 02:47:11 crc kubenswrapper[4912]: I1203 02:47:11.007537 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-8prl8_d65dbc95-9df7-49ec-9d0a-b883b70b4e3e/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:11 crc kubenswrapper[4912]: I1203 02:47:11.325337 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c73f5ab7-bc17-41fa-80af-55fbb2ebf0de/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 02:47:11 crc kubenswrapper[4912]: I1203 02:47:11.375201 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ba93637e-3299-4ead-926f-0e2f84c31df9/nova-metadata-log/0.log" Dec 03 02:47:11 crc kubenswrapper[4912]: I1203 02:47:11.499065 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23c2e51b-07f1-4b86-bc02-29969db68f38/nova-api-api/0.log" Dec 03 02:47:11 crc kubenswrapper[4912]: I1203 02:47:11.855112 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e835f70a-04f2-43d3-ba92-6a6d20216a7d/mysql-bootstrap/0.log" Dec 03 02:47:11 crc kubenswrapper[4912]: I1203 02:47:11.900068 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e6cbac08-d327-4238-8335-a69cfb3f71b4/nova-scheduler-scheduler/0.log" Dec 03 02:47:12 crc kubenswrapper[4912]: I1203 02:47:12.103439 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e835f70a-04f2-43d3-ba92-6a6d20216a7d/mysql-bootstrap/0.log" Dec 03 02:47:12 crc kubenswrapper[4912]: I1203 02:47:12.121954 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e835f70a-04f2-43d3-ba92-6a6d20216a7d/galera/0.log" Dec 03 02:47:12 crc kubenswrapper[4912]: I1203 02:47:12.390376 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9a1a55bf-bc0f-4004-b540-ead98e30d25a/mysql-bootstrap/0.log" Dec 03 02:47:12 crc kubenswrapper[4912]: I1203 02:47:12.531004 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9a1a55bf-bc0f-4004-b540-ead98e30d25a/mysql-bootstrap/0.log" Dec 03 02:47:12 crc kubenswrapper[4912]: I1203 02:47:12.586346 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9a1a55bf-bc0f-4004-b540-ead98e30d25a/galera/0.log" Dec 03 02:47:12 crc kubenswrapper[4912]: I1203 02:47:12.739956 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b1cf8990-bf85-4de8-8121-73c43a7eccd9/openstackclient/0.log" Dec 03 02:47:12 crc kubenswrapper[4912]: I1203 02:47:12.870906 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-gbbh7_dd4efef6-5ce3-444a-a464-1f18f7cc2db4/ovn-controller/0.log" Dec 03 02:47:13 crc kubenswrapper[4912]: I1203 02:47:13.075118 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-p7c9n_290c5de7-2853-45ef-9b03-99a0d301d51f/openstack-network-exporter/0.log" Dec 03 02:47:13 crc kubenswrapper[4912]: I1203 02:47:13.240832 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6lm7j_2c613cf4-0585-466f-8db2-4c4f4b4765f0/ovsdb-server-init/0.log" Dec 03 02:47:13 crc kubenswrapper[4912]: I1203 02:47:13.472727 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6lm7j_2c613cf4-0585-466f-8db2-4c4f4b4765f0/ovsdb-server-init/0.log" Dec 03 02:47:13 crc kubenswrapper[4912]: I1203 02:47:13.495206 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6lm7j_2c613cf4-0585-466f-8db2-4c4f4b4765f0/ovs-vswitchd/0.log" Dec 03 02:47:13 crc kubenswrapper[4912]: I1203 02:47:13.495988 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6lm7j_2c613cf4-0585-466f-8db2-4c4f4b4765f0/ovsdb-server/0.log" Dec 03 02:47:13 crc kubenswrapper[4912]: I1203 02:47:13.792924 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-49qvj_04c823be-fed9-4ffd-b41b-30115e049634/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:13 crc kubenswrapper[4912]: I1203 02:47:13.891782 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_197a0bff-4ebc-4581-bc0f-77d459500339/openstack-network-exporter/0.log" Dec 03 02:47:13 crc kubenswrapper[4912]: I1203 02:47:13.991201 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_197a0bff-4ebc-4581-bc0f-77d459500339/ovn-northd/0.log" Dec 03 02:47:14 crc kubenswrapper[4912]: I1203 02:47:14.098174 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026/openstack-network-exporter/0.log" Dec 03 02:47:14 crc kubenswrapper[4912]: I1203 02:47:14.208830 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ad3ecd9c-b9b1-41c9-bbc9-ef5674f9e026/ovsdbserver-nb/0.log" Dec 03 02:47:14 crc kubenswrapper[4912]: I1203 02:47:14.406663 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d0dca179-6a1f-41ad-bad5-1d193e2583c8/openstack-network-exporter/0.log" Dec 03 02:47:14 crc kubenswrapper[4912]: I1203 02:47:14.487603 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_d0dca179-6a1f-41ad-bad5-1d193e2583c8/ovsdbserver-sb/0.log" Dec 03 02:47:14 crc kubenswrapper[4912]: I1203 02:47:14.969716 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_ba93637e-3299-4ead-926f-0e2f84c31df9/nova-metadata-metadata/0.log" Dec 03 02:47:15 crc kubenswrapper[4912]: I1203 02:47:15.108320 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69d59dff98-dh972_7eb76b80-5bef-4a43-a3d0-098c3029fcb2/placement-api/0.log" Dec 03 02:47:15 crc kubenswrapper[4912]: I1203 02:47:15.233773 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69d59dff98-dh972_7eb76b80-5bef-4a43-a3d0-098c3029fcb2/placement-log/0.log" Dec 03 02:47:15 crc kubenswrapper[4912]: I1203 02:47:15.252680 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/init-config-reloader/0.log" Dec 03 02:47:15 crc kubenswrapper[4912]: I1203 02:47:15.410612 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/init-config-reloader/0.log" Dec 03 02:47:15 crc kubenswrapper[4912]: I1203 02:47:15.448812 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/thanos-sidecar/0.log" Dec 03 02:47:15 crc kubenswrapper[4912]: I1203 02:47:15.529412 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/config-reloader/0.log" Dec 03 02:47:15 crc kubenswrapper[4912]: I1203 02:47:15.577497 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_f78bc494-84ea-43ff-b3fc-0af2f10c43b4/prometheus/0.log" Dec 03 02:47:15 crc kubenswrapper[4912]: I1203 02:47:15.752621 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1435b2ef-93c1-41ac-a2bd-f9af839b7894/setup-container/0.log" Dec 03 02:47:15 crc kubenswrapper[4912]: I1203 02:47:15.915060 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1435b2ef-93c1-41ac-a2bd-f9af839b7894/setup-container/0.log" Dec 03 02:47:16 crc kubenswrapper[4912]: I1203 02:47:16.028915 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb7b3762-41c4-4935-8882-526576167e0a/setup-container/0.log" Dec 03 02:47:16 crc kubenswrapper[4912]: I1203 02:47:16.124298 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_1435b2ef-93c1-41ac-a2bd-f9af839b7894/rabbitmq/0.log" Dec 03 02:47:16 crc kubenswrapper[4912]: I1203 02:47:16.249661 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb7b3762-41c4-4935-8882-526576167e0a/setup-container/0.log" Dec 03 02:47:16 crc kubenswrapper[4912]: I1203 02:47:16.277768 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_eb7b3762-41c4-4935-8882-526576167e0a/rabbitmq/0.log" Dec 03 02:47:16 crc kubenswrapper[4912]: I1203 02:47:16.337755 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-kgmzh_4d2c5206-f0c2-4df9-9a99-345471f00226/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:16 crc kubenswrapper[4912]: I1203 02:47:16.527511 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-c4gkz_0bf1960f-0a29-4cf3-9f54-c4e2f4b8d784/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:16 crc kubenswrapper[4912]: I1203 02:47:16.571778 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:47:16 crc kubenswrapper[4912]: E1203 02:47:16.572065 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:47:16 crc kubenswrapper[4912]: I1203 02:47:16.686598 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-6kzw8_d31b18ef-e64f-47ac-9aad-401795f9e3f0/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:16 crc kubenswrapper[4912]: I1203 02:47:16.788953 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-nfq59_58be2429-a01c-42d9-b773-da6cae52f9b3/ssh-known-hosts-edpm-deployment/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.024203 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7ddd66b7bc-vwpcg_540979f8-98e6-40f8-9f82-1e7b924d984e/proxy-server/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.173509 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6wv8k_1556422b-130c-4bf1-8010-ee7c6ea5f662/swift-ring-rebalance/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.206332 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7ddd66b7bc-vwpcg_540979f8-98e6-40f8-9f82-1e7b924d984e/proxy-httpd/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.357587 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/account-auditor/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.457461 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/account-reaper/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.490228 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/account-replicator/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.525701 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/account-server/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.598805 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/container-auditor/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.777788 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/container-server/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.779048 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/container-replicator/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.818688 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/container-updater/0.log" Dec 03 02:47:17 crc kubenswrapper[4912]: I1203 02:47:17.821711 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-auditor/0.log" Dec 03 02:47:18 crc kubenswrapper[4912]: I1203 02:47:18.020698 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-expirer/0.log" Dec 03 02:47:18 crc kubenswrapper[4912]: I1203 02:47:18.056831 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-server/0.log" Dec 03 02:47:18 crc kubenswrapper[4912]: I1203 02:47:18.095730 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-updater/0.log" Dec 03 02:47:18 crc kubenswrapper[4912]: I1203 02:47:18.104411 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/object-replicator/0.log" Dec 03 02:47:18 crc kubenswrapper[4912]: I1203 02:47:18.270561 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/swift-recon-cron/0.log" Dec 03 02:47:18 crc kubenswrapper[4912]: I1203 02:47:18.273297 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_761ebfbd-144d-42cc-ac4f-09a0c49ec7bb/rsync/0.log" Dec 03 02:47:18 crc kubenswrapper[4912]: I1203 02:47:18.361222 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wl9ng_d6d6c301-1060-4d03-8e8e-6d776c912cc2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:18 crc kubenswrapper[4912]: I1203 02:47:18.764145 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-twwxm_e34cc78e-e946-4d45-a3da-962e308ad972/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:18 crc kubenswrapper[4912]: I1203 02:47:18.958060 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_deabdb62-423d-4c64-bbea-dd488b38ab79/test-operator-logs-container/0.log" Dec 03 02:47:19 crc kubenswrapper[4912]: I1203 02:47:19.275556 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-c75zj_078c0c55-2cdc-4164-af87-c60aced0babf/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 02:47:19 crc kubenswrapper[4912]: I1203 02:47:19.703364 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d5f2ea7f-cdc9-4573-8b16-6561c761f91f/tempest-tests-tempest-tests-runner/0.log" Dec 03 02:47:31 crc kubenswrapper[4912]: I1203 02:47:31.571823 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:47:31 crc kubenswrapper[4912]: E1203 02:47:31.576636 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:47:32 crc kubenswrapper[4912]: I1203 02:47:32.421498 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e909a03f-b885-46d4-a139-833410cf9ff1/memcached/0.log" Dec 03 02:47:45 crc kubenswrapper[4912]: I1203 02:47:45.571994 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:47:45 crc kubenswrapper[4912]: E1203 02:47:45.572920 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:47:51 crc kubenswrapper[4912]: I1203 02:47:51.553907 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-twd26_56cdea32-a9a4-410b-9a9c-c237746b0582/kube-rbac-proxy/0.log" Dec 03 02:47:51 crc kubenswrapper[4912]: I1203 02:47:51.582692 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-twd26_56cdea32-a9a4-410b-9a9c-c237746b0582/manager/0.log" Dec 03 02:47:51 crc kubenswrapper[4912]: I1203 02:47:51.753987 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-ds5mt_70a932ec-65ac-4616-870b-9297bd5d0a28/kube-rbac-proxy/0.log" Dec 03 02:47:51 crc kubenswrapper[4912]: I1203 02:47:51.797437 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-ds5mt_70a932ec-65ac-4616-870b-9297bd5d0a28/manager/0.log" Dec 03 02:47:51 crc kubenswrapper[4912]: I1203 02:47:51.922841 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-tqphq_5118703e-2fdb-41d6-a76e-ab5d3f6adb95/kube-rbac-proxy/0.log" Dec 03 02:47:51 crc kubenswrapper[4912]: I1203 02:47:51.969647 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-tqphq_5118703e-2fdb-41d6-a76e-ab5d3f6adb95/manager/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.034633 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/util/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.232067 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/pull/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.252165 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/util/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.273028 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/pull/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.459717 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/pull/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.463035 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/util/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.482520 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_e6e262cee69b5b397b9f11713585338b282ff4d6176110f23ae0fb3bcc8xjj7_f67e2276-c381-45a2-a5e8-c4a353fd8244/extract/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.673718 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-g4n8g_d2643e7f-8db2-4374-b0d9-05a2130795aa/kube-rbac-proxy/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.717520 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jskq6_1df68f98-9a63-43ca-8e47-5c34c1e52581/kube-rbac-proxy/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.821087 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-g4n8g_d2643e7f-8db2-4374-b0d9-05a2130795aa/manager/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.975085 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-jskq6_1df68f98-9a63-43ca-8e47-5c34c1e52581/manager/0.log" Dec 03 02:47:52 crc kubenswrapper[4912]: I1203 02:47:52.997333 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-rb4z7_27605b80-0a7f-4b6b-9729-d0b6eaa0a74b/kube-rbac-proxy/0.log" Dec 03 02:47:53 crc kubenswrapper[4912]: I1203 02:47:53.039237 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-rb4z7_27605b80-0a7f-4b6b-9729-d0b6eaa0a74b/manager/0.log" Dec 03 02:47:53 crc kubenswrapper[4912]: I1203 02:47:53.195646 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-lzff2_045ab366-9392-40f9-94f5-95c983f65176/kube-rbac-proxy/0.log" Dec 03 02:47:53 crc kubenswrapper[4912]: I1203 02:47:53.394356 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-lzff2_045ab366-9392-40f9-94f5-95c983f65176/manager/0.log" Dec 03 02:47:53 crc kubenswrapper[4912]: I1203 02:47:53.612537 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-626ss_7b515331-ceab-4f45-9880-719c72dfcc4c/kube-rbac-proxy/0.log" Dec 03 02:47:53 crc kubenswrapper[4912]: I1203 02:47:53.649654 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-626ss_7b515331-ceab-4f45-9880-719c72dfcc4c/manager/0.log" Dec 03 02:47:53 crc kubenswrapper[4912]: I1203 02:47:53.780700 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-txtqb_9b23d447-fdc8-47a1-af38-a569b6ad2a8f/kube-rbac-proxy/0.log" Dec 03 02:47:53 crc kubenswrapper[4912]: I1203 02:47:53.909530 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-txtqb_9b23d447-fdc8-47a1-af38-a569b6ad2a8f/manager/0.log" Dec 03 02:47:53 crc kubenswrapper[4912]: I1203 02:47:53.953939 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-fxhh2_8ddf2fc1-446c-4a26-987a-f67ec4c51455/kube-rbac-proxy/0.log" Dec 03 02:47:54 crc kubenswrapper[4912]: I1203 02:47:54.023471 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-fxhh2_8ddf2fc1-446c-4a26-987a-f67ec4c51455/manager/0.log" Dec 03 02:47:54 crc kubenswrapper[4912]: I1203 02:47:54.155680 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-rjqjr_a7db7741-5afc-46bc-84c4-71c233ca9dde/kube-rbac-proxy/0.log" Dec 03 02:47:54 crc kubenswrapper[4912]: I1203 02:47:54.185682 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-rjqjr_a7db7741-5afc-46bc-84c4-71c233ca9dde/manager/0.log" Dec 03 02:47:54 crc kubenswrapper[4912]: I1203 02:47:54.507887 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-mjphw_6848c014-0383-482f-9d46-232cc8fafbe1/kube-rbac-proxy/0.log" Dec 03 02:47:54 crc kubenswrapper[4912]: I1203 02:47:54.517728 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-mjphw_6848c014-0383-482f-9d46-232cc8fafbe1/manager/0.log" Dec 03 02:47:54 crc kubenswrapper[4912]: I1203 02:47:54.674680 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-bdgdj_de517f42-de01-4dde-ac22-1a651a38926b/kube-rbac-proxy/0.log" Dec 03 02:47:54 crc kubenswrapper[4912]: I1203 02:47:54.765846 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-bdgdj_de517f42-de01-4dde-ac22-1a651a38926b/manager/0.log" Dec 03 02:47:54 crc kubenswrapper[4912]: I1203 02:47:54.812374 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dtgrm_baaa7298-1635-4de5-af87-68db95102082/kube-rbac-proxy/0.log" Dec 03 02:47:54 crc kubenswrapper[4912]: I1203 02:47:54.894510 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-dtgrm_baaa7298-1635-4de5-af87-68db95102082/manager/0.log" Dec 03 02:47:55 crc kubenswrapper[4912]: I1203 02:47:55.028157 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb_3e85fc81-3fbd-4af2-8c98-89892a4f3579/kube-rbac-proxy/0.log" Dec 03 02:47:55 crc kubenswrapper[4912]: I1203 02:47:55.038529 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4jg6xb_3e85fc81-3fbd-4af2-8c98-89892a4f3579/manager/0.log" Dec 03 02:47:55 crc kubenswrapper[4912]: I1203 02:47:55.494823 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-c796f8559-s567r_e96cfd80-56de-4ab0-98b0-8bdcb324e498/operator/0.log" Dec 03 02:47:55 crc kubenswrapper[4912]: I1203 02:47:55.545994 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-4nzrg_5d286a0d-a818-4cb6-a798-135fda9d947e/registry-server/0.log" Dec 03 02:47:55 crc kubenswrapper[4912]: I1203 02:47:55.781467 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-vw7dk_fd3cb745-7c4b-484b-93a5-7638dcd4d738/kube-rbac-proxy/0.log" Dec 03 02:47:55 crc kubenswrapper[4912]: I1203 02:47:55.845955 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-vw7dk_fd3cb745-7c4b-484b-93a5-7638dcd4d738/manager/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.053894 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-5n67v_7ec29496-d0ac-42a2-acea-7037dfde41fb/kube-rbac-proxy/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.065575 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-5n67v_7ec29496-d0ac-42a2-acea-7037dfde41fb/manager/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.138909 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-llhvt_b54a05e7-9cfd-401d-8e5d-87c8ed5ad321/operator/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.347224 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-8x84c_2ab18e2a-3358-44e6-ac55-80b072619e73/kube-rbac-proxy/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.453239 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-8x84c_2ab18e2a-3358-44e6-ac55-80b072619e73/manager/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.556688 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d66d5c957-xgv9t_82b74254-de20-48a4-bfae-21afed7954cd/kube-rbac-proxy/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.789555 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-7rstc_b5544f39-6c3a-4737-a8e3-389a6e66ba40/kube-rbac-proxy/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.792992 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-7rstc_b5544f39-6c3a-4737-a8e3-389a6e66ba40/manager/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.924177 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86464467f8-sq576_10227b7f-ff98-4f4a-862f-3ef78b4e7737/manager/0.log" Dec 03 02:47:56 crc kubenswrapper[4912]: I1203 02:47:56.970461 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d66d5c957-xgv9t_82b74254-de20-48a4-bfae-21afed7954cd/manager/0.log" Dec 03 02:47:57 crc kubenswrapper[4912]: I1203 02:47:57.020720 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-jqzw7_ea0ec30d-2c13-49fd-883b-767024ff632c/kube-rbac-proxy/0.log" Dec 03 02:47:57 crc kubenswrapper[4912]: I1203 02:47:57.043290 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-jqzw7_ea0ec30d-2c13-49fd-883b-767024ff632c/manager/0.log" Dec 03 02:47:58 crc kubenswrapper[4912]: I1203 02:47:58.571600 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:47:58 crc kubenswrapper[4912]: E1203 02:47:58.571971 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:48:09 crc kubenswrapper[4912]: I1203 02:48:09.572321 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:48:09 crc kubenswrapper[4912]: E1203 02:48:09.573233 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:48:18 crc kubenswrapper[4912]: I1203 02:48:18.648298 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-g5h6h_cff6f92d-904b-43ba-a0d7-2bae917cdcd8/control-plane-machine-set-operator/0.log" Dec 03 02:48:18 crc kubenswrapper[4912]: I1203 02:48:18.766192 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tqgnx_ac6ae3d3-462a-4a82-a4a6-e225f12ce343/kube-rbac-proxy/0.log" Dec 03 02:48:18 crc kubenswrapper[4912]: I1203 02:48:18.789543 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tqgnx_ac6ae3d3-462a-4a82-a4a6-e225f12ce343/machine-api-operator/0.log" Dec 03 02:48:20 crc kubenswrapper[4912]: I1203 02:48:20.571966 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:48:20 crc kubenswrapper[4912]: E1203 02:48:20.572721 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:48:32 crc kubenswrapper[4912]: I1203 02:48:32.572254 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:48:32 crc kubenswrapper[4912]: E1203 02:48:32.573068 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:48:33 crc kubenswrapper[4912]: I1203 02:48:33.778932 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-9hkxj_52072628-05d2-4d43-8fea-11ec571a1a5f/cert-manager-controller/0.log" Dec 03 02:48:33 crc kubenswrapper[4912]: I1203 02:48:33.823150 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-dlkbd_2620530c-1377-4da3-a983-a0211b396c18/cert-manager-cainjector/0.log" Dec 03 02:48:33 crc kubenswrapper[4912]: I1203 02:48:33.885726 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-s6b5l_77d9d66c-7ec2-4cc0-8240-bd02f4b710c7/cert-manager-webhook/0.log" Dec 03 02:48:44 crc kubenswrapper[4912]: I1203 02:48:44.586492 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:48:44 crc kubenswrapper[4912]: E1203 02:48:44.587296 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:48:48 crc kubenswrapper[4912]: I1203 02:48:48.252333 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-7dtpc_0c9a6fef-57f2-4add-99ff-2807455dd07d/nmstate-console-plugin/0.log" Dec 03 02:48:48 crc kubenswrapper[4912]: I1203 02:48:48.424973 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-667c7_af9b5997-afe3-45f8-8f6e-1b4b427c41cd/nmstate-handler/0.log" Dec 03 02:48:48 crc kubenswrapper[4912]: I1203 02:48:48.479102 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-w95jw_ff447976-7803-4a25-8491-f2094e165f86/kube-rbac-proxy/0.log" Dec 03 02:48:48 crc kubenswrapper[4912]: I1203 02:48:48.485290 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-w95jw_ff447976-7803-4a25-8491-f2094e165f86/nmstate-metrics/0.log" Dec 03 02:48:48 crc kubenswrapper[4912]: I1203 02:48:48.687195 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-mxpqr_48b78ad2-74d3-43ff-a69d-e729d3082199/nmstate-operator/0.log" Dec 03 02:48:48 crc kubenswrapper[4912]: I1203 02:48:48.710898 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-wt4wl_df373e95-5be1-45c8-a5a0-58dfd53caf3c/nmstate-webhook/0.log" Dec 03 02:48:58 crc kubenswrapper[4912]: I1203 02:48:58.572110 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:48:58 crc kubenswrapper[4912]: E1203 02:48:58.574034 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:49:02 crc kubenswrapper[4912]: I1203 02:49:02.350557 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-669f4ccbc6-wkghn_32b6d349-0f43-45f8-93d6-c5f9f66273b2/kube-rbac-proxy/0.log" Dec 03 02:49:02 crc kubenswrapper[4912]: I1203 02:49:02.376548 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-669f4ccbc6-wkghn_32b6d349-0f43-45f8-93d6-c5f9f66273b2/manager/0.log" Dec 03 02:49:13 crc kubenswrapper[4912]: I1203 02:49:13.571480 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:49:13 crc kubenswrapper[4912]: E1203 02:49:13.572139 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:49:18 crc kubenswrapper[4912]: I1203 02:49:18.234209 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-kxsph_fe13031d-b775-459e-ad64-7445ca8cbe63/cluster-logging-operator/0.log" Dec 03 02:49:18 crc kubenswrapper[4912]: I1203 02:49:18.363707 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-dccc6_5d5921d6-eb7c-4430-b71f-104dc4fdff22/collector/0.log" Dec 03 02:49:18 crc kubenswrapper[4912]: I1203 02:49:18.409025 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_2ead2e0e-573f-4ef4-8122-34c91de2fabe/loki-compactor/0.log" Dec 03 02:49:18 crc kubenswrapper[4912]: I1203 02:49:18.604317 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-cr4z6_3aa94604-5389-4ef5-8e97-63167d5fb009/loki-distributor/0.log" Dec 03 02:49:18 crc kubenswrapper[4912]: I1203 02:49:18.679631 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-777c467b8d-tfg7h_00d2fcad-0e8b-4daf-ae2d-084456cd7144/gateway/0.log" Dec 03 02:49:18 crc kubenswrapper[4912]: I1203 02:49:18.722860 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-777c467b8d-tfg7h_00d2fcad-0e8b-4daf-ae2d-084456cd7144/opa/0.log" Dec 03 02:49:18 crc kubenswrapper[4912]: I1203 02:49:18.806652 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-777c467b8d-zsrn8_dd4ddd07-754f-48f9-b082-97b948a8d041/gateway/0.log" Dec 03 02:49:18 crc kubenswrapper[4912]: I1203 02:49:18.881678 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-777c467b8d-zsrn8_dd4ddd07-754f-48f9-b082-97b948a8d041/opa/0.log" Dec 03 02:49:18 crc kubenswrapper[4912]: I1203 02:49:18.933882 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_dfc8d0a7-a087-4dfb-a460-e63aeb64828e/loki-index-gateway/0.log" Dec 03 02:49:19 crc kubenswrapper[4912]: I1203 02:49:19.114887 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_33e70879-0df5-4529-969f-e88309e22cb0/loki-ingester/0.log" Dec 03 02:49:19 crc kubenswrapper[4912]: I1203 02:49:19.236979 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-vxq5n_3828ecda-60a7-4376-a5ec-f2730b95945d/loki-querier/0.log" Dec 03 02:49:19 crc kubenswrapper[4912]: I1203 02:49:19.319421 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-ljkc8_faaba064-9cc2-44fe-8516-29b31630da75/loki-query-frontend/0.log" Dec 03 02:49:25 crc kubenswrapper[4912]: I1203 02:49:25.572082 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:49:26 crc kubenswrapper[4912]: I1203 02:49:26.740136 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"59cd79276a11af231a9ec9908010de4aeb4e067da562d7b21d068e5a4c76b113"} Dec 03 02:49:35 crc kubenswrapper[4912]: I1203 02:49:35.511645 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-sqxw4_b4d86772-d2f0-4a24-80e7-339798fc2d2c/kube-rbac-proxy/0.log" Dec 03 02:49:35 crc kubenswrapper[4912]: I1203 02:49:35.555136 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-sqxw4_b4d86772-d2f0-4a24-80e7-339798fc2d2c/controller/0.log" Dec 03 02:49:35 crc kubenswrapper[4912]: I1203 02:49:35.655238 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-frr-files/0.log" Dec 03 02:49:35 crc kubenswrapper[4912]: I1203 02:49:35.889776 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-frr-files/0.log" Dec 03 02:49:35 crc kubenswrapper[4912]: I1203 02:49:35.897604 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-reloader/0.log" Dec 03 02:49:35 crc kubenswrapper[4912]: I1203 02:49:35.898829 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-metrics/0.log" Dec 03 02:49:35 crc kubenswrapper[4912]: I1203 02:49:35.937015 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-reloader/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.072685 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-frr-files/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.078892 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-reloader/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.091357 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-metrics/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.149538 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-metrics/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.298335 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-frr-files/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.329029 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-reloader/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.364506 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/cp-metrics/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.381630 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/controller/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.528608 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/frr-metrics/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.598592 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/kube-rbac-proxy/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.598662 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/kube-rbac-proxy-frr/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.719733 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/reloader/0.log" Dec 03 02:49:36 crc kubenswrapper[4912]: I1203 02:49:36.866971 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-jlww2_9f8130d5-88a6-4a62-a381-e98a3207d9b3/frr-k8s-webhook-server/0.log" Dec 03 02:49:37 crc kubenswrapper[4912]: I1203 02:49:37.106303 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-76f7d69d98-sl5x6_ae8cc701-2530-4ef7-bd0c-596cb88ffc3a/manager/0.log" Dec 03 02:49:37 crc kubenswrapper[4912]: I1203 02:49:37.188941 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-58487cf9c5-5g5xb_cb334b7c-36eb-4863-a2ad-2f4f126a9354/webhook-server/0.log" Dec 03 02:49:37 crc kubenswrapper[4912]: I1203 02:49:37.370608 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9nt4p_bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36/kube-rbac-proxy/0.log" Dec 03 02:49:38 crc kubenswrapper[4912]: I1203 02:49:38.197504 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9nt4p_bd9f5e2d-164a-4a09-8cf4-57f1ecd84e36/speaker/0.log" Dec 03 02:49:38 crc kubenswrapper[4912]: I1203 02:49:38.834006 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8c4wr_1ccecd71-076b-4f63-830a-ad1507c443b8/frr/0.log" Dec 03 02:49:52 crc kubenswrapper[4912]: I1203 02:49:52.367247 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/util/0.log" Dec 03 02:49:52 crc kubenswrapper[4912]: I1203 02:49:52.509163 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/util/0.log" Dec 03 02:49:52 crc kubenswrapper[4912]: I1203 02:49:52.546082 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/pull/0.log" Dec 03 02:49:52 crc kubenswrapper[4912]: I1203 02:49:52.575290 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/pull/0.log" Dec 03 02:49:52 crc kubenswrapper[4912]: I1203 02:49:52.737480 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/util/0.log" Dec 03 02:49:52 crc kubenswrapper[4912]: I1203 02:49:52.761609 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/extract/0.log" Dec 03 02:49:52 crc kubenswrapper[4912]: I1203 02:49:52.770389 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8vjlhh_25cacc35-89f1-4f0e-a621-ee002ce1f7e0/pull/0.log" Dec 03 02:49:52 crc kubenswrapper[4912]: I1203 02:49:52.935053 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/util/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.084861 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/util/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.111580 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/pull/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.129803 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/pull/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.309553 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/util/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.340590 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/pull/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.351382 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212frkd55_e58f04a9-b86f-4988-98bb-b15d2b753399/extract/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.523953 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/util/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.669360 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/pull/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.701384 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/util/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.728322 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/pull/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.897193 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/pull/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.914115 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/util/0.log" Dec 03 02:49:53 crc kubenswrapper[4912]: I1203 02:49:53.987545 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210vhlkc_e3e0e84c-fd2b-41ac-adda-0c32d431e5f3/extract/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.106830 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/util/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.301373 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/pull/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.340296 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/util/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.360546 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/pull/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.565190 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/pull/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.601913 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/util/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.615248 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463f4z4bf_26654d73-149b-418b-8c5e-00653837545b/extract/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.777376 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/util/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.967091 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/pull/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.976330 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/util/0.log" Dec 03 02:49:54 crc kubenswrapper[4912]: I1203 02:49:54.998133 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/pull/0.log" Dec 03 02:49:55 crc kubenswrapper[4912]: I1203 02:49:55.154806 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/util/0.log" Dec 03 02:49:55 crc kubenswrapper[4912]: I1203 02:49:55.177216 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/extract/0.log" Dec 03 02:49:55 crc kubenswrapper[4912]: I1203 02:49:55.180049 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83nrf9k_8e6614d3-319c-4f5f-8937-fbe619a1c55a/pull/0.log" Dec 03 02:49:55 crc kubenswrapper[4912]: I1203 02:49:55.332016 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-utilities/0.log" Dec 03 02:49:55 crc kubenswrapper[4912]: I1203 02:49:55.528577 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-utilities/0.log" Dec 03 02:49:55 crc kubenswrapper[4912]: I1203 02:49:55.543060 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-content/0.log" Dec 03 02:49:55 crc kubenswrapper[4912]: I1203 02:49:55.585000 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-content/0.log" Dec 03 02:49:55 crc kubenswrapper[4912]: I1203 02:49:55.768729 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-utilities/0.log" Dec 03 02:49:55 crc kubenswrapper[4912]: I1203 02:49:55.837955 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/extract-content/0.log" Dec 03 02:49:56 crc kubenswrapper[4912]: I1203 02:49:56.023636 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-utilities/0.log" Dec 03 02:49:56 crc kubenswrapper[4912]: I1203 02:49:56.163249 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-content/0.log" Dec 03 02:49:56 crc kubenswrapper[4912]: I1203 02:49:56.203931 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-content/0.log" Dec 03 02:49:56 crc kubenswrapper[4912]: I1203 02:49:56.218896 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-utilities/0.log" Dec 03 02:49:56 crc kubenswrapper[4912]: I1203 02:49:56.484312 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-utilities/0.log" Dec 03 02:49:56 crc kubenswrapper[4912]: I1203 02:49:56.553201 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/extract-content/0.log" Dec 03 02:49:56 crc kubenswrapper[4912]: I1203 02:49:56.748617 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-24t69_f8dce103-6de5-4159-a1e3-4ae68c513ee7/marketplace-operator/0.log" Dec 03 02:49:56 crc kubenswrapper[4912]: I1203 02:49:56.918729 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-utilities/0.log" Dec 03 02:49:56 crc kubenswrapper[4912]: I1203 02:49:56.981282 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-utilities/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.055740 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-content/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.075815 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jfl9b_d18182ed-8455-4575-aa3d-99fa4e26f604/registry-server/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.178233 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-content/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.427683 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-utilities/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.429989 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/extract-content/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.628046 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-utilities/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.832219 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-qqv59_5aaca572-9fa6-4c8f-8563-0609714513b0/registry-server/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.893636 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-content/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.901595 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-utilities/0.log" Dec 03 02:49:57 crc kubenswrapper[4912]: I1203 02:49:57.948940 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-content/0.log" Dec 03 02:49:58 crc kubenswrapper[4912]: I1203 02:49:58.028284 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bkzm2_8e9fe039-8cbd-4a92-9b85-4a17818ade89/registry-server/0.log" Dec 03 02:49:58 crc kubenswrapper[4912]: I1203 02:49:58.124390 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-content/0.log" Dec 03 02:49:58 crc kubenswrapper[4912]: I1203 02:49:58.125865 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/extract-utilities/0.log" Dec 03 02:49:58 crc kubenswrapper[4912]: I1203 02:49:58.704290 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pxmmk_c67383be-5842-4d3c-9a66-ce5c7075a721/registry-server/0.log" Dec 03 02:50:12 crc kubenswrapper[4912]: I1203 02:50:12.336046 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-spkm8_377b37d8-070e-4463-9ea8-c698c4e289a4/prometheus-operator/0.log" Dec 03 02:50:12 crc kubenswrapper[4912]: I1203 02:50:12.540592 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-84fc64b9db-hmqf5_194f5401-404a-43b8-b826-3df930390630/prometheus-operator-admission-webhook/0.log" Dec 03 02:50:12 crc kubenswrapper[4912]: I1203 02:50:12.606695 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-84fc64b9db-x8jr2_45c9685c-8c5e-4da5-9c6f-d1f00f6d83e0/prometheus-operator-admission-webhook/0.log" Dec 03 02:50:12 crc kubenswrapper[4912]: I1203 02:50:12.767514 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-bsscx_f4a067c7-ef55-4185-8315-a61b5bfc8e22/operator/0.log" Dec 03 02:50:12 crc kubenswrapper[4912]: I1203 02:50:12.833307 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-96rq5_9d7ffd59-b442-43e2-b7b4-5a2fc42507d3/observability-ui-dashboards/0.log" Dec 03 02:50:12 crc kubenswrapper[4912]: I1203 02:50:12.947839 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-trfdr_1bba9178-9a01-47ac-9657-c277217019ca/perses-operator/0.log" Dec 03 02:50:27 crc kubenswrapper[4912]: I1203 02:50:27.697275 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-669f4ccbc6-wkghn_32b6d349-0f43-45f8-93d6-c5f9f66273b2/manager/0.log" Dec 03 02:50:27 crc kubenswrapper[4912]: I1203 02:50:27.726147 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-669f4ccbc6-wkghn_32b6d349-0f43-45f8-93d6-c5f9f66273b2/kube-rbac-proxy/0.log" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.332351 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k82cj"] Dec 03 02:51:12 crc kubenswrapper[4912]: E1203 02:51:12.335787 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerName="registry-server" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.335824 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerName="registry-server" Dec 03 02:51:12 crc kubenswrapper[4912]: E1203 02:51:12.336084 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerName="extract-content" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.336112 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerName="extract-content" Dec 03 02:51:12 crc kubenswrapper[4912]: E1203 02:51:12.336143 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerName="registry-server" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.336154 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerName="registry-server" Dec 03 02:51:12 crc kubenswrapper[4912]: E1203 02:51:12.336175 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerName="extract-content" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.336186 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerName="extract-content" Dec 03 02:51:12 crc kubenswrapper[4912]: E1203 02:51:12.336215 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerName="extract-utilities" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.336227 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerName="extract-utilities" Dec 03 02:51:12 crc kubenswrapper[4912]: E1203 02:51:12.336250 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerName="extract-utilities" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.336263 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerName="extract-utilities" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.336876 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b8caf38-5e5c-4f03-976b-45d693b3dcf8" containerName="registry-server" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.336951 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="35140842-fcec-49dc-a3a2-3ecfad78e611" containerName="registry-server" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.341348 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.344133 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k82cj"] Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.435150 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ncbf\" (UniqueName: \"kubernetes.io/projected/7fc98c87-5380-448f-a82f-a9ab33315176-kube-api-access-2ncbf\") pod \"certified-operators-k82cj\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.435783 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-utilities\") pod \"certified-operators-k82cj\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.435955 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-catalog-content\") pod \"certified-operators-k82cj\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.538274 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-utilities\") pod \"certified-operators-k82cj\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.538336 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-catalog-content\") pod \"certified-operators-k82cj\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.538533 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ncbf\" (UniqueName: \"kubernetes.io/projected/7fc98c87-5380-448f-a82f-a9ab33315176-kube-api-access-2ncbf\") pod \"certified-operators-k82cj\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.539025 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-utilities\") pod \"certified-operators-k82cj\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.539058 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-catalog-content\") pod \"certified-operators-k82cj\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.566039 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ncbf\" (UniqueName: \"kubernetes.io/projected/7fc98c87-5380-448f-a82f-a9ab33315176-kube-api-access-2ncbf\") pod \"certified-operators-k82cj\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:12 crc kubenswrapper[4912]: I1203 02:51:12.675392 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:13 crc kubenswrapper[4912]: I1203 02:51:13.522654 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k82cj"] Dec 03 02:51:13 crc kubenswrapper[4912]: W1203 02:51:13.528684 4912 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fc98c87_5380_448f_a82f_a9ab33315176.slice/crio-ad9af561385f2836d2fabc2b86147a4b29fa9d66d59f5bc4e99556867615fbb6 WatchSource:0}: Error finding container ad9af561385f2836d2fabc2b86147a4b29fa9d66d59f5bc4e99556867615fbb6: Status 404 returned error can't find the container with id ad9af561385f2836d2fabc2b86147a4b29fa9d66d59f5bc4e99556867615fbb6 Dec 03 02:51:14 crc kubenswrapper[4912]: I1203 02:51:14.081738 4912 generic.go:334] "Generic (PLEG): container finished" podID="7fc98c87-5380-448f-a82f-a9ab33315176" containerID="a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29" exitCode=0 Dec 03 02:51:14 crc kubenswrapper[4912]: I1203 02:51:14.081857 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82cj" event={"ID":"7fc98c87-5380-448f-a82f-a9ab33315176","Type":"ContainerDied","Data":"a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29"} Dec 03 02:51:14 crc kubenswrapper[4912]: I1203 02:51:14.082174 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82cj" event={"ID":"7fc98c87-5380-448f-a82f-a9ab33315176","Type":"ContainerStarted","Data":"ad9af561385f2836d2fabc2b86147a4b29fa9d66d59f5bc4e99556867615fbb6"} Dec 03 02:51:15 crc kubenswrapper[4912]: I1203 02:51:15.092539 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82cj" event={"ID":"7fc98c87-5380-448f-a82f-a9ab33315176","Type":"ContainerStarted","Data":"cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33"} Dec 03 02:51:16 crc kubenswrapper[4912]: I1203 02:51:16.107254 4912 generic.go:334] "Generic (PLEG): container finished" podID="7fc98c87-5380-448f-a82f-a9ab33315176" containerID="cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33" exitCode=0 Dec 03 02:51:16 crc kubenswrapper[4912]: I1203 02:51:16.107326 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82cj" event={"ID":"7fc98c87-5380-448f-a82f-a9ab33315176","Type":"ContainerDied","Data":"cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33"} Dec 03 02:51:17 crc kubenswrapper[4912]: I1203 02:51:17.120435 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82cj" event={"ID":"7fc98c87-5380-448f-a82f-a9ab33315176","Type":"ContainerStarted","Data":"7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f"} Dec 03 02:51:17 crc kubenswrapper[4912]: I1203 02:51:17.150228 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k82cj" podStartSLOduration=2.658699764 podStartE2EDuration="5.150198058s" podCreationTimestamp="2025-12-03 02:51:12 +0000 UTC" firstStartedPulling="2025-12-03 02:51:14.084723534 +0000 UTC m=+8859.726744114" lastFinishedPulling="2025-12-03 02:51:16.576221808 +0000 UTC m=+8862.218242408" observedRunningTime="2025-12-03 02:51:17.143197391 +0000 UTC m=+8862.785218001" watchObservedRunningTime="2025-12-03 02:51:17.150198058 +0000 UTC m=+8862.792218658" Dec 03 02:51:22 crc kubenswrapper[4912]: I1203 02:51:22.676550 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:22 crc kubenswrapper[4912]: I1203 02:51:22.677673 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:22 crc kubenswrapper[4912]: I1203 02:51:22.725358 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:23 crc kubenswrapper[4912]: I1203 02:51:23.281322 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:23 crc kubenswrapper[4912]: I1203 02:51:23.366530 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k82cj"] Dec 03 02:51:25 crc kubenswrapper[4912]: I1203 02:51:25.220207 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k82cj" podUID="7fc98c87-5380-448f-a82f-a9ab33315176" containerName="registry-server" containerID="cri-o://7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f" gracePeriod=2 Dec 03 02:51:25 crc kubenswrapper[4912]: I1203 02:51:25.805447 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:25 crc kubenswrapper[4912]: I1203 02:51:25.980678 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-catalog-content\") pod \"7fc98c87-5380-448f-a82f-a9ab33315176\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " Dec 03 02:51:25 crc kubenswrapper[4912]: I1203 02:51:25.980924 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-utilities\") pod \"7fc98c87-5380-448f-a82f-a9ab33315176\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " Dec 03 02:51:25 crc kubenswrapper[4912]: I1203 02:51:25.981024 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ncbf\" (UniqueName: \"kubernetes.io/projected/7fc98c87-5380-448f-a82f-a9ab33315176-kube-api-access-2ncbf\") pod \"7fc98c87-5380-448f-a82f-a9ab33315176\" (UID: \"7fc98c87-5380-448f-a82f-a9ab33315176\") " Dec 03 02:51:25 crc kubenswrapper[4912]: I1203 02:51:25.982273 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-utilities" (OuterVolumeSpecName: "utilities") pod "7fc98c87-5380-448f-a82f-a9ab33315176" (UID: "7fc98c87-5380-448f-a82f-a9ab33315176"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:51:25 crc kubenswrapper[4912]: I1203 02:51:25.988189 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc98c87-5380-448f-a82f-a9ab33315176-kube-api-access-2ncbf" (OuterVolumeSpecName: "kube-api-access-2ncbf") pod "7fc98c87-5380-448f-a82f-a9ab33315176" (UID: "7fc98c87-5380-448f-a82f-a9ab33315176"). InnerVolumeSpecName "kube-api-access-2ncbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.032239 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fc98c87-5380-448f-a82f-a9ab33315176" (UID: "7fc98c87-5380-448f-a82f-a9ab33315176"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.084074 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.084117 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ncbf\" (UniqueName: \"kubernetes.io/projected/7fc98c87-5380-448f-a82f-a9ab33315176-kube-api-access-2ncbf\") on node \"crc\" DevicePath \"\"" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.084130 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fc98c87-5380-448f-a82f-a9ab33315176-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.237722 4912 generic.go:334] "Generic (PLEG): container finished" podID="7fc98c87-5380-448f-a82f-a9ab33315176" containerID="7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f" exitCode=0 Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.238061 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82cj" event={"ID":"7fc98c87-5380-448f-a82f-a9ab33315176","Type":"ContainerDied","Data":"7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f"} Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.238089 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k82cj" event={"ID":"7fc98c87-5380-448f-a82f-a9ab33315176","Type":"ContainerDied","Data":"ad9af561385f2836d2fabc2b86147a4b29fa9d66d59f5bc4e99556867615fbb6"} Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.238110 4912 scope.go:117] "RemoveContainer" containerID="7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.238272 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k82cj" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.280897 4912 scope.go:117] "RemoveContainer" containerID="cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.285787 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k82cj"] Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.296203 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k82cj"] Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.306834 4912 scope.go:117] "RemoveContainer" containerID="a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.377404 4912 scope.go:117] "RemoveContainer" containerID="7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f" Dec 03 02:51:26 crc kubenswrapper[4912]: E1203 02:51:26.378000 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f\": container with ID starting with 7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f not found: ID does not exist" containerID="7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.378081 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f"} err="failed to get container status \"7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f\": rpc error: code = NotFound desc = could not find container \"7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f\": container with ID starting with 7a70a3ccca8ede0490676b6909a9519a408df9997c989a910b5ac32b3123b87f not found: ID does not exist" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.378135 4912 scope.go:117] "RemoveContainer" containerID="cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33" Dec 03 02:51:26 crc kubenswrapper[4912]: E1203 02:51:26.378542 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33\": container with ID starting with cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33 not found: ID does not exist" containerID="cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.378592 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33"} err="failed to get container status \"cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33\": rpc error: code = NotFound desc = could not find container \"cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33\": container with ID starting with cf9216f9cf1689fe6b3d78282395ac6cb8a5a6d25e723463f150f5cf88e5bb33 not found: ID does not exist" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.378623 4912 scope.go:117] "RemoveContainer" containerID="a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29" Dec 03 02:51:26 crc kubenswrapper[4912]: E1203 02:51:26.381800 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29\": container with ID starting with a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29 not found: ID does not exist" containerID="a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.381864 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29"} err="failed to get container status \"a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29\": rpc error: code = NotFound desc = could not find container \"a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29\": container with ID starting with a6807356f8b02ec2b3f97b55af6294078bb852a00cecd769e065b613a839de29 not found: ID does not exist" Dec 03 02:51:26 crc kubenswrapper[4912]: I1203 02:51:26.596848 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc98c87-5380-448f-a82f-a9ab33315176" path="/var/lib/kubelet/pods/7fc98c87-5380-448f-a82f-a9ab33315176/volumes" Dec 03 02:51:48 crc kubenswrapper[4912]: I1203 02:51:48.078414 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:51:48 crc kubenswrapper[4912]: I1203 02:51:48.079405 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:52:18 crc kubenswrapper[4912]: I1203 02:52:18.078548 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:52:18 crc kubenswrapper[4912]: I1203 02:52:18.079227 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:52:27 crc kubenswrapper[4912]: I1203 02:52:27.453623 4912 generic.go:334] "Generic (PLEG): container finished" podID="ed036cee-58c4-4577-8785-c1903431dd3a" containerID="5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25" exitCode=0 Dec 03 02:52:27 crc kubenswrapper[4912]: I1203 02:52:27.454266 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-x4djm/must-gather-5qw4w" event={"ID":"ed036cee-58c4-4577-8785-c1903431dd3a","Type":"ContainerDied","Data":"5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25"} Dec 03 02:52:27 crc kubenswrapper[4912]: I1203 02:52:27.455290 4912 scope.go:117] "RemoveContainer" containerID="5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25" Dec 03 02:52:27 crc kubenswrapper[4912]: I1203 02:52:27.855066 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x4djm_must-gather-5qw4w_ed036cee-58c4-4577-8785-c1903431dd3a/gather/0.log" Dec 03 02:52:31 crc kubenswrapper[4912]: I1203 02:52:31.152934 4912 scope.go:117] "RemoveContainer" containerID="b6a117927523a3ecf9be155eb08cf11ab9a062c34795de15506c089c9e072cbe" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.091261 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-x4djm/must-gather-5qw4w"] Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.092125 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-x4djm/must-gather-5qw4w" podUID="ed036cee-58c4-4577-8785-c1903431dd3a" containerName="copy" containerID="cri-o://1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe" gracePeriod=2 Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.109252 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-x4djm/must-gather-5qw4w"] Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.585414 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x4djm_must-gather-5qw4w_ed036cee-58c4-4577-8785-c1903431dd3a/copy/0.log" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.586328 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.627517 4912 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-x4djm_must-gather-5qw4w_ed036cee-58c4-4577-8785-c1903431dd3a/copy/0.log" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.628071 4912 generic.go:334] "Generic (PLEG): container finished" podID="ed036cee-58c4-4577-8785-c1903431dd3a" containerID="1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe" exitCode=143 Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.628117 4912 scope.go:117] "RemoveContainer" containerID="1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.628174 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-x4djm/must-gather-5qw4w" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.649683 4912 scope.go:117] "RemoveContainer" containerID="5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.691587 4912 scope.go:117] "RemoveContainer" containerID="1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe" Dec 03 02:52:41 crc kubenswrapper[4912]: E1203 02:52:41.692146 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe\": container with ID starting with 1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe not found: ID does not exist" containerID="1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.692181 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe"} err="failed to get container status \"1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe\": rpc error: code = NotFound desc = could not find container \"1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe\": container with ID starting with 1a2fd6929ab640e2faa1d77c2c44e6c1e53914ba2c58a5b79cbb09552ec624fe not found: ID does not exist" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.692201 4912 scope.go:117] "RemoveContainer" containerID="5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25" Dec 03 02:52:41 crc kubenswrapper[4912]: E1203 02:52:41.693198 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25\": container with ID starting with 5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25 not found: ID does not exist" containerID="5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.693238 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25"} err="failed to get container status \"5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25\": rpc error: code = NotFound desc = could not find container \"5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25\": container with ID starting with 5745d0c2be276e0b9a3578cd8b1ea441d59bc2448b0c873de7e107039ca62f25 not found: ID does not exist" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.728183 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed036cee-58c4-4577-8785-c1903431dd3a-must-gather-output\") pod \"ed036cee-58c4-4577-8785-c1903431dd3a\" (UID: \"ed036cee-58c4-4577-8785-c1903431dd3a\") " Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.728380 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2lhj\" (UniqueName: \"kubernetes.io/projected/ed036cee-58c4-4577-8785-c1903431dd3a-kube-api-access-f2lhj\") pod \"ed036cee-58c4-4577-8785-c1903431dd3a\" (UID: \"ed036cee-58c4-4577-8785-c1903431dd3a\") " Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.734963 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed036cee-58c4-4577-8785-c1903431dd3a-kube-api-access-f2lhj" (OuterVolumeSpecName: "kube-api-access-f2lhj") pod "ed036cee-58c4-4577-8785-c1903431dd3a" (UID: "ed036cee-58c4-4577-8785-c1903431dd3a"). InnerVolumeSpecName "kube-api-access-f2lhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.831590 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2lhj\" (UniqueName: \"kubernetes.io/projected/ed036cee-58c4-4577-8785-c1903431dd3a-kube-api-access-f2lhj\") on node \"crc\" DevicePath \"\"" Dec 03 02:52:41 crc kubenswrapper[4912]: I1203 02:52:41.936161 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed036cee-58c4-4577-8785-c1903431dd3a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ed036cee-58c4-4577-8785-c1903431dd3a" (UID: "ed036cee-58c4-4577-8785-c1903431dd3a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:52:42 crc kubenswrapper[4912]: I1203 02:52:42.035871 4912 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ed036cee-58c4-4577-8785-c1903431dd3a-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 02:52:42 crc kubenswrapper[4912]: I1203 02:52:42.588771 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed036cee-58c4-4577-8785-c1903431dd3a" path="/var/lib/kubelet/pods/ed036cee-58c4-4577-8785-c1903431dd3a/volumes" Dec 03 02:52:48 crc kubenswrapper[4912]: I1203 02:52:48.078242 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:52:48 crc kubenswrapper[4912]: I1203 02:52:48.080242 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:52:48 crc kubenswrapper[4912]: I1203 02:52:48.080477 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:52:48 crc kubenswrapper[4912]: I1203 02:52:48.081700 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"59cd79276a11af231a9ec9908010de4aeb4e067da562d7b21d068e5a4c76b113"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:52:48 crc kubenswrapper[4912]: I1203 02:52:48.082279 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://59cd79276a11af231a9ec9908010de4aeb4e067da562d7b21d068e5a4c76b113" gracePeriod=600 Dec 03 02:52:48 crc kubenswrapper[4912]: I1203 02:52:48.721892 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="59cd79276a11af231a9ec9908010de4aeb4e067da562d7b21d068e5a4c76b113" exitCode=0 Dec 03 02:52:48 crc kubenswrapper[4912]: I1203 02:52:48.721984 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"59cd79276a11af231a9ec9908010de4aeb4e067da562d7b21d068e5a4c76b113"} Dec 03 02:52:48 crc kubenswrapper[4912]: I1203 02:52:48.722306 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerStarted","Data":"b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108"} Dec 03 02:52:48 crc kubenswrapper[4912]: I1203 02:52:48.722340 4912 scope.go:117] "RemoveContainer" containerID="f2d8d241c49a081f51c42880621f72c4cc26d122201063e87f20ef023942f9ee" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.047337 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mhh9m"] Dec 03 02:53:23 crc kubenswrapper[4912]: E1203 02:53:23.049237 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed036cee-58c4-4577-8785-c1903431dd3a" containerName="copy" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.049304 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed036cee-58c4-4577-8785-c1903431dd3a" containerName="copy" Dec 03 02:53:23 crc kubenswrapper[4912]: E1203 02:53:23.049380 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc98c87-5380-448f-a82f-a9ab33315176" containerName="extract-content" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.049394 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc98c87-5380-448f-a82f-a9ab33315176" containerName="extract-content" Dec 03 02:53:23 crc kubenswrapper[4912]: E1203 02:53:23.049471 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc98c87-5380-448f-a82f-a9ab33315176" containerName="extract-utilities" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.049492 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc98c87-5380-448f-a82f-a9ab33315176" containerName="extract-utilities" Dec 03 02:53:23 crc kubenswrapper[4912]: E1203 02:53:23.049776 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc98c87-5380-448f-a82f-a9ab33315176" containerName="registry-server" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.049844 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc98c87-5380-448f-a82f-a9ab33315176" containerName="registry-server" Dec 03 02:53:23 crc kubenswrapper[4912]: E1203 02:53:23.049929 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed036cee-58c4-4577-8785-c1903431dd3a" containerName="gather" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.049946 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed036cee-58c4-4577-8785-c1903431dd3a" containerName="gather" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.051020 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc98c87-5380-448f-a82f-a9ab33315176" containerName="registry-server" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.051074 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed036cee-58c4-4577-8785-c1903431dd3a" containerName="copy" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.051118 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed036cee-58c4-4577-8785-c1903431dd3a" containerName="gather" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.057569 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.101267 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mhh9m"] Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.177763 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-catalog-content\") pod \"redhat-operators-mhh9m\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.177922 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qvkg\" (UniqueName: \"kubernetes.io/projected/dff45d72-974b-41eb-9734-c4c8a45927ea-kube-api-access-9qvkg\") pod \"redhat-operators-mhh9m\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.178123 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-utilities\") pod \"redhat-operators-mhh9m\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.281687 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-utilities\") pod \"redhat-operators-mhh9m\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.282115 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-catalog-content\") pod \"redhat-operators-mhh9m\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.282225 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-utilities\") pod \"redhat-operators-mhh9m\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.282280 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qvkg\" (UniqueName: \"kubernetes.io/projected/dff45d72-974b-41eb-9734-c4c8a45927ea-kube-api-access-9qvkg\") pod \"redhat-operators-mhh9m\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.282471 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-catalog-content\") pod \"redhat-operators-mhh9m\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.309550 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qvkg\" (UniqueName: \"kubernetes.io/projected/dff45d72-974b-41eb-9734-c4c8a45927ea-kube-api-access-9qvkg\") pod \"redhat-operators-mhh9m\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.403826 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:23 crc kubenswrapper[4912]: I1203 02:53:23.897755 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mhh9m"] Dec 03 02:53:24 crc kubenswrapper[4912]: I1203 02:53:24.225755 4912 generic.go:334] "Generic (PLEG): container finished" podID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerID="933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662" exitCode=0 Dec 03 02:53:24 crc kubenswrapper[4912]: I1203 02:53:24.225804 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mhh9m" event={"ID":"dff45d72-974b-41eb-9734-c4c8a45927ea","Type":"ContainerDied","Data":"933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662"} Dec 03 02:53:24 crc kubenswrapper[4912]: I1203 02:53:24.225858 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mhh9m" event={"ID":"dff45d72-974b-41eb-9734-c4c8a45927ea","Type":"ContainerStarted","Data":"db21c1ebd54994a617331f712014b56312a7e547d0a18d80360e30eb1714a004"} Dec 03 02:53:24 crc kubenswrapper[4912]: I1203 02:53:24.228540 4912 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 02:53:25 crc kubenswrapper[4912]: I1203 02:53:25.240552 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mhh9m" event={"ID":"dff45d72-974b-41eb-9734-c4c8a45927ea","Type":"ContainerStarted","Data":"072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3"} Dec 03 02:53:28 crc kubenswrapper[4912]: I1203 02:53:28.273096 4912 generic.go:334] "Generic (PLEG): container finished" podID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerID="072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3" exitCode=0 Dec 03 02:53:28 crc kubenswrapper[4912]: I1203 02:53:28.273149 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mhh9m" event={"ID":"dff45d72-974b-41eb-9734-c4c8a45927ea","Type":"ContainerDied","Data":"072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3"} Dec 03 02:53:31 crc kubenswrapper[4912]: I1203 02:53:31.279368 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mhh9m" event={"ID":"dff45d72-974b-41eb-9734-c4c8a45927ea","Type":"ContainerStarted","Data":"d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c"} Dec 03 02:53:31 crc kubenswrapper[4912]: I1203 02:53:31.308518 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mhh9m" podStartSLOduration=3.257476331 podStartE2EDuration="8.308498983s" podCreationTimestamp="2025-12-03 02:53:23 +0000 UTC" firstStartedPulling="2025-12-03 02:53:24.227957523 +0000 UTC m=+8989.869978093" lastFinishedPulling="2025-12-03 02:53:29.278980145 +0000 UTC m=+8994.921000745" observedRunningTime="2025-12-03 02:53:31.30279249 +0000 UTC m=+8996.944813060" watchObservedRunningTime="2025-12-03 02:53:31.308498983 +0000 UTC m=+8996.950519543" Dec 03 02:53:33 crc kubenswrapper[4912]: I1203 02:53:33.407792 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:33 crc kubenswrapper[4912]: I1203 02:53:33.408488 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:34 crc kubenswrapper[4912]: I1203 02:53:34.484244 4912 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mhh9m" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerName="registry-server" probeResult="failure" output=< Dec 03 02:53:34 crc kubenswrapper[4912]: timeout: failed to connect service ":50051" within 1s Dec 03 02:53:34 crc kubenswrapper[4912]: > Dec 03 02:53:43 crc kubenswrapper[4912]: I1203 02:53:43.475258 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:43 crc kubenswrapper[4912]: I1203 02:53:43.574563 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:43 crc kubenswrapper[4912]: I1203 02:53:43.724628 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mhh9m"] Dec 03 02:53:45 crc kubenswrapper[4912]: I1203 02:53:45.509537 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mhh9m" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerName="registry-server" containerID="cri-o://d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c" gracePeriod=2 Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.137467 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.316536 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-utilities\") pod \"dff45d72-974b-41eb-9734-c4c8a45927ea\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.316904 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-catalog-content\") pod \"dff45d72-974b-41eb-9734-c4c8a45927ea\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.316937 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qvkg\" (UniqueName: \"kubernetes.io/projected/dff45d72-974b-41eb-9734-c4c8a45927ea-kube-api-access-9qvkg\") pod \"dff45d72-974b-41eb-9734-c4c8a45927ea\" (UID: \"dff45d72-974b-41eb-9734-c4c8a45927ea\") " Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.317972 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-utilities" (OuterVolumeSpecName: "utilities") pod "dff45d72-974b-41eb-9734-c4c8a45927ea" (UID: "dff45d72-974b-41eb-9734-c4c8a45927ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.326020 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dff45d72-974b-41eb-9734-c4c8a45927ea-kube-api-access-9qvkg" (OuterVolumeSpecName: "kube-api-access-9qvkg") pod "dff45d72-974b-41eb-9734-c4c8a45927ea" (UID: "dff45d72-974b-41eb-9734-c4c8a45927ea"). InnerVolumeSpecName "kube-api-access-9qvkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.420169 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qvkg\" (UniqueName: \"kubernetes.io/projected/dff45d72-974b-41eb-9734-c4c8a45927ea-kube-api-access-9qvkg\") on node \"crc\" DevicePath \"\"" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.420212 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.441707 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dff45d72-974b-41eb-9734-c4c8a45927ea" (UID: "dff45d72-974b-41eb-9734-c4c8a45927ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.522550 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dff45d72-974b-41eb-9734-c4c8a45927ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.529640 4912 generic.go:334] "Generic (PLEG): container finished" podID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerID="d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c" exitCode=0 Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.529795 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mhh9m" event={"ID":"dff45d72-974b-41eb-9734-c4c8a45927ea","Type":"ContainerDied","Data":"d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c"} Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.529856 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mhh9m" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.529886 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mhh9m" event={"ID":"dff45d72-974b-41eb-9734-c4c8a45927ea","Type":"ContainerDied","Data":"db21c1ebd54994a617331f712014b56312a7e547d0a18d80360e30eb1714a004"} Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.529905 4912 scope.go:117] "RemoveContainer" containerID="d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.574797 4912 scope.go:117] "RemoveContainer" containerID="072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.628137 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mhh9m"] Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.630554 4912 scope.go:117] "RemoveContainer" containerID="933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.646886 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mhh9m"] Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.672082 4912 scope.go:117] "RemoveContainer" containerID="d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c" Dec 03 02:53:46 crc kubenswrapper[4912]: E1203 02:53:46.675068 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c\": container with ID starting with d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c not found: ID does not exist" containerID="d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.675120 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c"} err="failed to get container status \"d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c\": rpc error: code = NotFound desc = could not find container \"d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c\": container with ID starting with d793c62f66a84c10078b4a73df3d71fdfa2e01e2d0b2ec3b877f9602f011d62c not found: ID does not exist" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.675155 4912 scope.go:117] "RemoveContainer" containerID="072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3" Dec 03 02:53:46 crc kubenswrapper[4912]: E1203 02:53:46.675760 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3\": container with ID starting with 072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3 not found: ID does not exist" containerID="072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.676082 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3"} err="failed to get container status \"072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3\": rpc error: code = NotFound desc = could not find container \"072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3\": container with ID starting with 072ebdc6d7530eb8bb9dabf5375dbbffb7eb505eacfeeb2f6b82a3db0d886cc3 not found: ID does not exist" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.676302 4912 scope.go:117] "RemoveContainer" containerID="933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662" Dec 03 02:53:46 crc kubenswrapper[4912]: E1203 02:53:46.677072 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662\": container with ID starting with 933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662 not found: ID does not exist" containerID="933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662" Dec 03 02:53:46 crc kubenswrapper[4912]: I1203 02:53:46.677295 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662"} err="failed to get container status \"933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662\": rpc error: code = NotFound desc = could not find container \"933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662\": container with ID starting with 933f9b80d2a8979775aa7505314d643942dea56d63ae967e4055a827d8ba6662 not found: ID does not exist" Dec 03 02:53:48 crc kubenswrapper[4912]: I1203 02:53:48.592076 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" path="/var/lib/kubelet/pods/dff45d72-974b-41eb-9734-c4c8a45927ea/volumes" Dec 03 02:54:48 crc kubenswrapper[4912]: I1203 02:54:48.079724 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:54:48 crc kubenswrapper[4912]: I1203 02:54:48.081228 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:55:18 crc kubenswrapper[4912]: I1203 02:55:18.077720 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:55:18 crc kubenswrapper[4912]: I1203 02:55:18.078449 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:55:48 crc kubenswrapper[4912]: I1203 02:55:48.078278 4912 patch_prober.go:28] interesting pod/machine-config-daemon-2qk89 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 02:55:48 crc kubenswrapper[4912]: I1203 02:55:48.079161 4912 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 02:55:48 crc kubenswrapper[4912]: I1203 02:55:48.079233 4912 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" Dec 03 02:55:48 crc kubenswrapper[4912]: I1203 02:55:48.081486 4912 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108"} pod="openshift-machine-config-operator/machine-config-daemon-2qk89" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 02:55:48 crc kubenswrapper[4912]: I1203 02:55:48.081609 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerName="machine-config-daemon" containerID="cri-o://b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" gracePeriod=600 Dec 03 02:55:48 crc kubenswrapper[4912]: E1203 02:55:48.220696 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:55:48 crc kubenswrapper[4912]: I1203 02:55:48.270359 4912 generic.go:334] "Generic (PLEG): container finished" podID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" exitCode=0 Dec 03 02:55:48 crc kubenswrapper[4912]: I1203 02:55:48.270455 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" event={"ID":"92796847-a53f-4ffd-8746-7fd65c0ee4ea","Type":"ContainerDied","Data":"b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108"} Dec 03 02:55:48 crc kubenswrapper[4912]: I1203 02:55:48.270508 4912 scope.go:117] "RemoveContainer" containerID="59cd79276a11af231a9ec9908010de4aeb4e067da562d7b21d068e5a4c76b113" Dec 03 02:55:48 crc kubenswrapper[4912]: I1203 02:55:48.271686 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:55:48 crc kubenswrapper[4912]: E1203 02:55:48.272375 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:56:00 crc kubenswrapper[4912]: I1203 02:56:00.572376 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:56:00 crc kubenswrapper[4912]: E1203 02:56:00.573912 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:56:11 crc kubenswrapper[4912]: I1203 02:56:11.572776 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:56:11 crc kubenswrapper[4912]: E1203 02:56:11.573559 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:56:25 crc kubenswrapper[4912]: I1203 02:56:25.571988 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:56:25 crc kubenswrapper[4912]: E1203 02:56:25.572748 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:56:39 crc kubenswrapper[4912]: I1203 02:56:39.572263 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:56:39 crc kubenswrapper[4912]: E1203 02:56:39.573333 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:56:52 crc kubenswrapper[4912]: I1203 02:56:52.572504 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:56:52 crc kubenswrapper[4912]: E1203 02:56:52.573576 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:57:04 crc kubenswrapper[4912]: I1203 02:57:04.596787 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:57:04 crc kubenswrapper[4912]: E1203 02:57:04.599527 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:57:18 crc kubenswrapper[4912]: I1203 02:57:18.573190 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:57:18 crc kubenswrapper[4912]: E1203 02:57:18.574126 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.005304 4912 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2trbd"] Dec 03 02:57:30 crc kubenswrapper[4912]: E1203 02:57:30.006314 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerName="registry-server" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.006334 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerName="registry-server" Dec 03 02:57:30 crc kubenswrapper[4912]: E1203 02:57:30.006361 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerName="extract-content" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.006367 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerName="extract-content" Dec 03 02:57:30 crc kubenswrapper[4912]: E1203 02:57:30.006386 4912 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerName="extract-utilities" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.006392 4912 state_mem.go:107] "Deleted CPUSet assignment" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerName="extract-utilities" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.006631 4912 memory_manager.go:354] "RemoveStaleState removing state" podUID="dff45d72-974b-41eb-9734-c4c8a45927ea" containerName="registry-server" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.008415 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.030565 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2trbd"] Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.161576 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-catalog-content\") pod \"community-operators-2trbd\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.161799 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-utilities\") pod \"community-operators-2trbd\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.162008 4912 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njwwf\" (UniqueName: \"kubernetes.io/projected/7d244a41-2252-4c76-883b-56fd7592779f-kube-api-access-njwwf\") pod \"community-operators-2trbd\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.264593 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-catalog-content\") pod \"community-operators-2trbd\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.264709 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-utilities\") pod \"community-operators-2trbd\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.264753 4912 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njwwf\" (UniqueName: \"kubernetes.io/projected/7d244a41-2252-4c76-883b-56fd7592779f-kube-api-access-njwwf\") pod \"community-operators-2trbd\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.265143 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-catalog-content\") pod \"community-operators-2trbd\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.265426 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-utilities\") pod \"community-operators-2trbd\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.284724 4912 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njwwf\" (UniqueName: \"kubernetes.io/projected/7d244a41-2252-4c76-883b-56fd7592779f-kube-api-access-njwwf\") pod \"community-operators-2trbd\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.329882 4912 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:30 crc kubenswrapper[4912]: I1203 02:57:30.846515 4912 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2trbd"] Dec 03 02:57:31 crc kubenswrapper[4912]: I1203 02:57:31.577787 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d244a41-2252-4c76-883b-56fd7592779f" containerID="bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7" exitCode=0 Dec 03 02:57:31 crc kubenswrapper[4912]: I1203 02:57:31.578019 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2trbd" event={"ID":"7d244a41-2252-4c76-883b-56fd7592779f","Type":"ContainerDied","Data":"bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7"} Dec 03 02:57:31 crc kubenswrapper[4912]: I1203 02:57:31.578040 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2trbd" event={"ID":"7d244a41-2252-4c76-883b-56fd7592779f","Type":"ContainerStarted","Data":"381c654f565c72bfce8d768af9b2be98310c715172d03d6f891e4237db9beb46"} Dec 03 02:57:32 crc kubenswrapper[4912]: I1203 02:57:32.571756 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:57:32 crc kubenswrapper[4912]: E1203 02:57:32.572468 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:57:33 crc kubenswrapper[4912]: I1203 02:57:33.622496 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2trbd" event={"ID":"7d244a41-2252-4c76-883b-56fd7592779f","Type":"ContainerStarted","Data":"b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4"} Dec 03 02:57:34 crc kubenswrapper[4912]: I1203 02:57:34.646981 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d244a41-2252-4c76-883b-56fd7592779f" containerID="b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4" exitCode=0 Dec 03 02:57:34 crc kubenswrapper[4912]: I1203 02:57:34.647341 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2trbd" event={"ID":"7d244a41-2252-4c76-883b-56fd7592779f","Type":"ContainerDied","Data":"b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4"} Dec 03 02:57:35 crc kubenswrapper[4912]: I1203 02:57:35.658583 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2trbd" event={"ID":"7d244a41-2252-4c76-883b-56fd7592779f","Type":"ContainerStarted","Data":"d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b"} Dec 03 02:57:35 crc kubenswrapper[4912]: I1203 02:57:35.686015 4912 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2trbd" podStartSLOduration=3.205557972 podStartE2EDuration="6.685995093s" podCreationTimestamp="2025-12-03 02:57:29 +0000 UTC" firstStartedPulling="2025-12-03 02:57:31.583680348 +0000 UTC m=+9237.225700908" lastFinishedPulling="2025-12-03 02:57:35.064117469 +0000 UTC m=+9240.706138029" observedRunningTime="2025-12-03 02:57:35.67987994 +0000 UTC m=+9241.321900510" watchObservedRunningTime="2025-12-03 02:57:35.685995093 +0000 UTC m=+9241.328015663" Dec 03 02:57:40 crc kubenswrapper[4912]: I1203 02:57:40.330598 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:40 crc kubenswrapper[4912]: I1203 02:57:40.331276 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:40 crc kubenswrapper[4912]: I1203 02:57:40.406417 4912 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:40 crc kubenswrapper[4912]: I1203 02:57:40.811869 4912 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:40 crc kubenswrapper[4912]: I1203 02:57:40.894890 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2trbd"] Dec 03 02:57:42 crc kubenswrapper[4912]: I1203 02:57:42.737249 4912 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2trbd" podUID="7d244a41-2252-4c76-883b-56fd7592779f" containerName="registry-server" containerID="cri-o://d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b" gracePeriod=2 Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.282280 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.383593 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-catalog-content\") pod \"7d244a41-2252-4c76-883b-56fd7592779f\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.383695 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-utilities\") pod \"7d244a41-2252-4c76-883b-56fd7592779f\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.383807 4912 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njwwf\" (UniqueName: \"kubernetes.io/projected/7d244a41-2252-4c76-883b-56fd7592779f-kube-api-access-njwwf\") pod \"7d244a41-2252-4c76-883b-56fd7592779f\" (UID: \"7d244a41-2252-4c76-883b-56fd7592779f\") " Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.384768 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-utilities" (OuterVolumeSpecName: "utilities") pod "7d244a41-2252-4c76-883b-56fd7592779f" (UID: "7d244a41-2252-4c76-883b-56fd7592779f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.390463 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d244a41-2252-4c76-883b-56fd7592779f-kube-api-access-njwwf" (OuterVolumeSpecName: "kube-api-access-njwwf") pod "7d244a41-2252-4c76-883b-56fd7592779f" (UID: "7d244a41-2252-4c76-883b-56fd7592779f"). InnerVolumeSpecName "kube-api-access-njwwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.448040 4912 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d244a41-2252-4c76-883b-56fd7592779f" (UID: "7d244a41-2252-4c76-883b-56fd7592779f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.487234 4912 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.487303 4912 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d244a41-2252-4c76-883b-56fd7592779f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.487327 4912 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njwwf\" (UniqueName: \"kubernetes.io/projected/7d244a41-2252-4c76-883b-56fd7592779f-kube-api-access-njwwf\") on node \"crc\" DevicePath \"\"" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.754729 4912 generic.go:334] "Generic (PLEG): container finished" podID="7d244a41-2252-4c76-883b-56fd7592779f" containerID="d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b" exitCode=0 Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.754792 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2trbd" event={"ID":"7d244a41-2252-4c76-883b-56fd7592779f","Type":"ContainerDied","Data":"d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b"} Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.754838 4912 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2trbd" event={"ID":"7d244a41-2252-4c76-883b-56fd7592779f","Type":"ContainerDied","Data":"381c654f565c72bfce8d768af9b2be98310c715172d03d6f891e4237db9beb46"} Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.754870 4912 scope.go:117] "RemoveContainer" containerID="d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.754944 4912 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2trbd" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.795408 4912 scope.go:117] "RemoveContainer" containerID="b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.821419 4912 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2trbd"] Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.835489 4912 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2trbd"] Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.855510 4912 scope.go:117] "RemoveContainer" containerID="bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.896545 4912 scope.go:117] "RemoveContainer" containerID="d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b" Dec 03 02:57:43 crc kubenswrapper[4912]: E1203 02:57:43.897084 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b\": container with ID starting with d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b not found: ID does not exist" containerID="d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.897135 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b"} err="failed to get container status \"d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b\": rpc error: code = NotFound desc = could not find container \"d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b\": container with ID starting with d0262050d338a79012334ceed78dfdf4513e62d85c2cd4d4b02223f586656f0b not found: ID does not exist" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.897168 4912 scope.go:117] "RemoveContainer" containerID="b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4" Dec 03 02:57:43 crc kubenswrapper[4912]: E1203 02:57:43.897956 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4\": container with ID starting with b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4 not found: ID does not exist" containerID="b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.898016 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4"} err="failed to get container status \"b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4\": rpc error: code = NotFound desc = could not find container \"b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4\": container with ID starting with b226d816f4f13cc3eda235d4e03ca64b2274fe147970267951ea42dc3a398cd4 not found: ID does not exist" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.898045 4912 scope.go:117] "RemoveContainer" containerID="bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7" Dec 03 02:57:43 crc kubenswrapper[4912]: E1203 02:57:43.898405 4912 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7\": container with ID starting with bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7 not found: ID does not exist" containerID="bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7" Dec 03 02:57:43 crc kubenswrapper[4912]: I1203 02:57:43.898559 4912 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7"} err="failed to get container status \"bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7\": rpc error: code = NotFound desc = could not find container \"bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7\": container with ID starting with bfe343c3fedaea8e0b283048e6e23cbf5d8146d135c03795bae0799d7599bbb7 not found: ID does not exist" Dec 03 02:57:44 crc kubenswrapper[4912]: I1203 02:57:44.577020 4912 scope.go:117] "RemoveContainer" containerID="b1148083da92c5df5a78d316fee1557e835996a906d66a09a80a540431d67108" Dec 03 02:57:44 crc kubenswrapper[4912]: E1203 02:57:44.577859 4912 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2qk89_openshift-machine-config-operator(92796847-a53f-4ffd-8746-7fd65c0ee4ea)\"" pod="openshift-machine-config-operator/machine-config-daemon-2qk89" podUID="92796847-a53f-4ffd-8746-7fd65c0ee4ea" Dec 03 02:57:44 crc kubenswrapper[4912]: I1203 02:57:44.612022 4912 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d244a41-2252-4c76-883b-56fd7592779f" path="/var/lib/kubelet/pods/7d244a41-2252-4c76-883b-56fd7592779f/volumes"